Next Article in Journal
Pollutant Removal Efficiency of Pilot-Scale Horizontal Subsurface Flow Constructed Wetlands Treating Landfill Leachate
Next Article in Special Issue
Quantum Neural Networks Approach for Water Discharge Forecast
Previous Article in Journal
Intelligent Scheduling of a Pulsating Assembly Flow Shop Considering a Multifunctional Automated Guided Vehicle
Previous Article in Special Issue
Tensor Decomposition Through Neural Architectures
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Neural Network Method of Controllers’ Parametric Optimization with Variable Structure and Semi-Permanent Integration Based on the Computation of Second-Order Sensitivity Functions

by
Serhii Vladov
1,
Lukasz Scislo
2,*,
Nina Szczepanik-Ścisło
3,4,
Anatoliy Sachenko
5,6 and
Victoria Vysotska
7
1
Kharkiv National University of Internal Affairs, 27, L. Landau Avenue, 61080 Kharkiv, Ukraine
2
Faculty of Electrical and Computer Engineering, Cracow University of Technology, 24, Warszawska, 31-155 Cracow, Poland
3
Faculty of Environmental Engineering and Energy, Cracow University of Technology, 24, Warszawska, 31-155 Cracow, Poland
4
CERN, European Organization for Nuclear Research, 1, Esplanade des Particules, 1211 Meyrin, Switzerland
5
Research Institute for Intelligent Computer Systems, West Ukrainian National University, 11, Lvivska Street, 46009 Ternopil, Ukraine
6
Department of Teleinformatics, Casimir Pulaski Radom University, 29, Malczewskiego Street, 26-600 Radom, Poland
7
Information Systems and Networks Department, Lviv Polytechnic National University, 12, Bandera Street, 79013 Lviv, Ukraine
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(5), 2586; https://doi.org/10.3390/app15052586
Submission received: 20 January 2025 / Revised: 25 February 2025 / Accepted: 26 February 2025 / Published: 27 February 2025
(This article belongs to the Special Issue Innovations in Artificial Neural Network Applications)

Abstract

:
This article presents a method for researching processes in automatic control systems based on the operator approach for modelling the control object and the controller. Within the method framework, a system of equations has been developed that describes the relations between the control error, the reference and control action, the output coordinate and the controller and the control object operators. The traditional PI controller modification, including a switching function for adaptation to operating conditions, allows for the system’s effective control in real time. The controller optimization algorithm is based on a functional expression with weighting coefficients that take into account control errors and the control action. To train the neural network through implementing the proposed method, a multilayer architecture was used, including nonlinear activation functions and a dynamic training rate, which ensure high accuracy and accelerated convergence. The TV3-117 turboshaft engine was chosen as the research object, which allows the method to be demonstrated in practical applications in aviation technology. The experimental results showed a significant improvement in control characteristics, including a reduction in the gas-generator rotor speed parameter transient time to ≈1, which is two times faster than the traditional method, where the transient process reaches ≈0.5. The model achieved a maximum accuracy of 0.993 with 160 training epochs, minimizing the error function to 0.005. In comparison with similar approaches, the proposed method demonstrated better results in accuracy and training speed, which was confirmed by a reduction in the number of iterations by 1.36 times and an improvement in the mean square error by 1.86–6.02 times.

1. Introduction and Related Works

Modern automatic control systems (ACSs) [1,2] are becoming increasingly complex due to increasing requirements for their reliability, accuracy and adaptability. One of the key tasks is to optimize the controllers’ parameters to ensure the transient processes’ stability under varying external conditions. Of particular interest are controllers with a variable structure and semi-permanent integration [3,4], which allow the system parameters’ flexible adaptation depending on its current state.
Modern approaches [5,6,7] to the parametric optimization of variable-structure controllers are based on the analysis and numerical calculations of the mathematical methods used. Key areas of research are gradient optimization methods [8,9], approaches based on sensitivity theory [10,11] and the optimal control methods used [12,13]. For example, in [14,15], the research is devoted to the use of gradient algorithms, and attention is primarily paid to the computation of the quality functional gradient in ACSs using first-order sensitivity. The key drawbacks of modern approaches to parametric optimization are the first-order sensitivity limiting its applications, which reduces the tuning accuracy in systems with pronounced nonlinearity [14,15]. In addition, gradient methods often require significant computational resources, especially when working with high-dimensional systems, which limits their practical implementation [8,16]. Optimal control and sensitivity theory methods also face scalability problems and difficulties in being integrated with real-time adaptive controllers [10,11,12,13].
The indicated shortcomings [5,6,7,8,9,10,11,12,13,14,15,16] are eliminated in methods based on the computation of second-order sensitivity functions. Methods based on the computation of second-order sensitivity functions [16,17,18] allow for increasing the parametric tuning accuracy. Methods based on the computation of second-order sensitivity provide more accurate parametric tuning by taking into account the interaction between parameters and their influence on system changes [16,17]. These methods take into account changes in the system, not only depending on the current state but also on the influence of parameter interactions [16,17]. However, such approaches are often difficult to implement due to the high computational load and the need to use specialized software [16,18]. This limitation can be mitigated by using neural network technologies capable of efficiently processing data in large amounts. However, the need for specialized software tools makes integration with modern neural network architectures difficult.
Therefore, a significant direction is the development of an adaptive neural network controller, which changes the parameters that are capable of changing in real time depending on the control object dynamics [19,20,21,22,23,24]. However, most existing approaches [19,20] are limited to using static switching functions, which do not allow for the system’s nonlinear behaviour to be fully considered. This reduces adaptation efficiency in complex transient conditions [20] or causes extensive time delays [21]. Known neural network controllers are developed based on the following neural network architectures: RBF networks [20], Jordan neural networks [21], Elman neural networks [22], Hopfield neural networks [23] and Hamming neural networks [24]. Known neural network controllers’ key drawback is their limited ability to integrate sensitivity functions that are sensitive to changes in the control object’s dynamics. For example, RBF networks do not adapt effectively enough to rapidly changing conditions, and Elman and Jordan neural networks require significant computing resources to account for complex nonlinear dynamics in real time. Thus, the issue of integrating sensitivity functions with modern neural network technologies has not been sufficiently overcome.
Despite the success in the application of second-order sensitivity methods, the task of developing algorithms that would be researched effectively under extensive delay conditions, strong nonlinearities and high uncertainty remains relevant. In addition, most research [5,8,9,10,13,14,16,19,20] is aimed at analyzing stable control objects, which limits the application of existing methods to systems with changing characteristics.
Another unsolved problem is the selection and optimization of the controller switching function. The influence of different function options on the quality of transient processes has not been researched sufficiently [6,7,12,15,17], which limits the controllers’ use in dynamically changing conditions.
The search-free method for a controller’s parametric adaptation described in [25,26,27] has a number of advantages, including the use of controller adaptive parameters, which improves the system’s tuning accuracy and takes into account the interactions between the parameters. Regularization and gradient procedures with the inclusion of a dynamic training rate make this method especially useful for systems with variable structures. The main disadvantage of the search-free method [25,26,27,28] is its limited ability to take into account strong nonlinearities and a control object’s extensive time delay. The high computational complexity associated with the analytical calculation of second-order sensitivity functions complicates its application to systems with a variable structure [5,6] and dynamically changing characteristics [7]. Table 1 provides a comparison of the approaches considered.
To improve the search-free method [25,26,27,28], it is necessary to integrate neural network approaches for automated modelling and the numerical computation of second-order sensitivity functions, which will reduce the computational complexity and increase accuracy. It is vital to take into account the nonlinearities and the control object’s delays by implementing adaptive switching functions in the controller operator. It is also necessary to automate the adaptation of controller parameters using error backpropagation methods modified for dynamic systems and implement a dynamic training rate mechanism to accelerate convergence. To improve the optimization stability, regularization should be included in the computation of the Hessian matrix and the optimality criterion, which will prevent overtraining and enhance system reliability in uncertain and noisy environments.
Thus, this research aims to develop a neural network method for variable-structure controllers with semi-permanent integration parametric optimization, taking into account nonlinearities, delays and high uncertainty, to improve the accuracy and stability of the ACS. The research object is an ACS with variable-structure controllers and semi-permanent integration. The research subject is parametric optimization methods for controllers based on calculating second-order sensitivity functions and using neural network approaches to take into account the control object’s nonlinear and dynamic characteristics. This research’s main contribution is the development of a neural network method for variable-structure controllers with the semi-permanent integration of parametric optimization, which takes into account nonlinearities, delays and high uncertainty. This allows for improving the ACS’s accuracy and stability, increasing its adaptability in the face of complex transient processes and objects’ dynamically changing characteristics.

2. Materials and Methods

According to [25,26,27,28], delays are often encountered in industrial processes such as transportation and substance mixing. Provided that the ratio τ o b T o b m > 1 (where τob is the delay time and T o b m is the object’s maximum time constant), standard controllers such as P, PI and PID do not provide the required quality in transient processes. One of the control methods involves PI controllers with a delay, which can be used in the structure. Semi-constant integration in the PI controller is also used, which allows for avoiding an increase in the number of adjustable parameters. These controllers with a variable structure are difficult to adjust, which requires approximate methods, reducing their efficiency. In this regard, algorithmic tuning methods are necessary. Positive results from the gradient algorithm applied based on sensitivity theory [10,11] to calculate the proportional controller’s adjustable parameters confirm its applicability for solving the parametric optimization issue, which consists of minimizing a given optimization criterion. The parametric optimization issue is to find values for the controller’s adjustable parameters, vector θ = (θ1, θ2), so that the minimum optimization criterion is met. Thus, according to [10,11,25,26,27,28], the processes occurring in the system are described by the following equation:
ε ( t ,   θ ¯ ) = r ( t ) x ( t ) ;   u ( t ) = G c ( s ,   θ ¯ )   ·   ε ( t ,   θ ¯ ) ;   x ( t ) = G p ( p )   ·   u ( t ) ,
where ε(t, θ ¯ ) is the control system error; r(t) is the reference action; u(t) is the control action; x(t) is the ACS output coordinate; Gc(s, θ ¯ ) is the controller operator; Gp(s) is the control object operator; and s = d d t is the differentiation operator (according to [25,28]).
The control object operator, Gp(s), in this research is chosen in such a form that it is possible to describe the most complex dynamic object processes:
G p s = k o b T o b 1 · s + 1 · T o b 2 · s + 1 · e τ o b · s + f n o n l i n e x , u ,
where kob is the static gain; T o b 1 and T o b 2 are the time constants; τob is the delay time, and fnonline(x, u) is the nonlinear part depending on the system’s state.
The proposed controller operator Gc(s, θ ¯ ), which is a traditional PI controller extension including a switching function, has the form:
G c s = k p · 1 + 1 T i · s · g t , θ ¯ ,
where kp is the proportionality coefficient that determines the controller’s response intensity to a deviation from the set value, Ti is the integration time that characterizes the influence of previous errors on the controller’s action, and g(t, θ ¯ ) is the switching function that depends on time t and the parameter θ ¯ , which is responsible for the controller’s dynamic adaptation.
The function g(t, θ ¯ ) regulates changes in the controller’s behaviour depending on the system’s current operating conditions, based on [25], and it is represented by one of the following possible expressions:
g t , θ ¯ = k 1 , ε t , θ ¯ · ε ˙ t , θ ¯ 0 , k 2 · e α · t + k 3 , ε t , θ ¯ · ε ˙ t , θ ¯ < 0 ,
where k1, k2, k3, and α are adjustable parameters depending on θ.
The switching function g(t) in (4) is chosen to ensure the adaptive behaviour of the controller in a dynamically changing control environment [8], which is confirmed by both theoretical and empirical considerations. On the one hand, the presence of adjustable parameters (k1, k2, k3, α) allows for fine-tuning the controller’s gain, where the exponential factor e α · t reduces the influence of sharp disturbances, preventing overgain and oscillations, which ensures an optimal balance between the system’s speed and stability. On the other hand, empirical data on the gas-generator part of the TV3-117 engine operation’s modelling [13] demonstrate that the use of g(t) allows for reducing the transient process time by half compared to traditional methods and improves the optimization algorithm’s convergence, reducing the number iterations and minimizing the loss function to MSE ≈ 0.005. Experimental research [8,13,25,26,27,28] confirms that the function’s selected form ensures the controller’s reliable adaptation in real time, the system parameters’ correct restoration, and a smoother and more predictable transient process, which is confirmed by the Hessian matrix’s positive definiteness and the stability of the system’s dynamics.
The object operator parameters’ fundamental values (2) are taken, taking into account the significant delay condition, i.e., τ o b T o b m > 1 .
Taking into account the expression that determines the control action, u(t), values, similarly to [25], the analytical expression describing u(t) is presented as follows:
u t = θ 1 · ε t , θ ¯ + θ 2 s · ε t , θ ¯ , g t , ε t , θ ¯ , ε ˙ t , θ ¯ = 1 , θ 1 · ε t , θ ¯ + θ 2 s · 0 t , g t , ε t , θ ¯ , ε ˙ t , θ ¯ = 0 .
Similarly to [25], this research sets the task of investigating the influence of the different variants of the switching function (4) in the controller (5) on the transient processes’ quality. Figure 1 shows the controller’s structural diagram under research [25].
As a criterion for optimizing the automatic parametric optimization-generated algorithm, based on [29], the functionality widely used in the practice of automatic regulation was selected:
J = 0 a 1 · ε 2 t , θ ¯ + a 2 · u 2 t d t ,
where a1 and a2 are weighting coefficients.
In the formation of the automatic parametric optimization algorithm in this research, sensitivity theory methods [10,11] for a system’s continuous class were used [18,19] since the control action, u(t), discontinuity point’s coordinate tm = (1, 2, …, m) does not depend on θ ¯ [21,25,26].
Similarly to [25,26,27,28], in this research, a model was obtained for the computation of the controller’s adjustable parameters’ (5) sensitivity functions:
ξ i t = G p s · u t θ i ,   i = 1,2 ,
u t θ i = G c s , q ¯ θ i · ε t , θ ¯ ξ i t · G c s , θ ¯ ; u t θ 1 = ε t , θ ¯ ξ i t · G c s , θ ¯ ;
u t θ 2 = 1 p · ε t , θ ¯ ξ 2 t · G c s , θ ¯ , g t , ε t , θ ¯ , ε ˙ t , θ ¯ = 1 , 1 p · 0 t ξ 2 t · G c s , θ ¯ , g t , ε t , θ ¯ , ε ˙ t , θ ¯ = 0 .
During the optimization process, the adjustable parameter vector θ ¯ is changed in accordance with the gradient procedure [25] (Figure 2):
θ ¯ l = θ ¯ l 1 η l · I · θ ¯ I ε t , θ ¯ l 1 ,   l = 1,2 , ,
where Γ = diag(γ) = {δij, γij} is the weight coefficient diagonal matrix; δij is the Kronecker delta; θ ¯ I ε t , θ ¯ l 1 is the gradient vector (6); and ηl is the dynamic training rate depending on the change in J. The definition of the gradient vector θ ¯ I ε t , θ ¯ l 1 in this research is associated with the computation of the sensitivity function vector Ξ for the variable-structure controller under consideration:
Ξ = ( ξ 1 ( t ) , ξ 2 ( t ) ) , Ξ t = ξ 1 t θ ¯ 1 , ξ 2 t θ ¯ 2 , 2 Ξ t 2 = ξ 1 2 t θ ¯ 1 2 , ξ 2 2 t θ ¯ 2 2 .
The implementation of the automatic parametric optimization-generated algorithm on a computer requires stopping the formation of the optimization process; in [25,28,30,31], the following condition was applied and tested for the automatic parametric optimization algorithms based on sensitivity theory:
S j l = S J l + 1 + 1   at   J l × J l 1 < 0 , 0   at   J l × J l 1 > 0 ,
S J j l = S J j l + 1 + 1   at   J θ ¯ l θ ¯ j × J θ ¯ l 1 θ ¯ j < 0 , 0   at   J θ ¯ l θ ¯ j × J θ ¯ l 1 θ ¯ j > 0 , j = 1,2 ¯ .
where J l = J θ ¯ l J θ ¯ l 1 is the difference between the optimality criteria in the next two iterations.
The automatic parametric optimization-generated algorithm for computation-adjustable parameters is considered complete when the following condition is met:
S J l n S J S J 1 n J 1 l S J 2 n J 2 l ,
where nSJ, n J 1 , and n J 2 are given positive values characterizing the corresponding number of changes in the sign of the optimality criterion increments and the gradient components of each sign. The specific values of the parameters nSJ, n J 1 , and n J 2 were obtained as a preliminary research result.
For the automatic parametric optimization-generated algorithm, in the next stage, the analytical dependencies were determined to achieve the research aim of checking the reliability of the adjustable parameters θ ¯ * computed based on criterion (6) for the minimum value.
In this research, the Hessian matrix H J = 2 J θ ¯ 2 + λ · J was calculated for this aim, where λ is the regularization coefficient, and the computational technique from [25,26,27,28] is additionally applied. As is known, a positive definite matrix, H(J), at the point θ ¯ * indicates that the minimum for criterion (6) has been found:
H J = 2 J θ 1 2 + λ · J 2 J θ 1 θ 2 + λ · J 2 J θ 2 θ 1 + λ · J 2 J θ 2 2 + λ · J = 2 J θ i θ j + λ · J ,   i , j = 1,2 ¯ .
Sylvester’s criterion for a positive definite matrix (15) is presented as follows:
2 J θ 1 2 + λ · J > 0 ,   2 J θ 2 2 + λ · J > 0 ,   d e t H J > 0 .
The definition of the elements in matrix (15) is based on the second-order sensitivity functions; that is,
ξ 1 t θ 1 ξ 1 t θ 2 ξ 2 t θ 1 ξ 2 t θ 2 = ξ i j t ,   i , j = 1,2 ¯ .
The second-order sensitivity function characterizes a change’s effect on the change rate of another sensitivity function, allowing one to take into account the system parameters’ interactions under nonlinear dynamics [16,17]. This approach provides a more in-depth analysis of the parameters’ influence on the system’s behaviour than first-order methods and allows one to more accurately adjust the controller’s settings. This is especially important for systems with significant delays and complex nonlinearities, where traditional optimization methods may not be effective enough [16]. Empirical research has confirmed that the second-order sensitivity functions used lead to accelerated convergence of the optimization algorithm and increased system stability [18].
The analytical search for sensitivity functions (15) is a labour-intensive task [8,12,14]. Therefore, in this research, a more straightforward method based on the sensitivity model [16,17,18,25,26] was used. Within this research framework, similarly to [25], a second-order sensitivity function model was obtained for system (1) with controller (5):
ξ i j t = G p s · 2 u t θ i θ j ,   i , j = 1,2 ¯ , 2 u t θ i θ j = 2 G c s , θ ¯ θ i θ j · ε t , θ ¯ G c s , θ ¯ θ i · ξ j t G c s , θ ¯ θ j · ξ i t G c s , θ ¯ · ξ i j t , 2 u t θ i θ j = 2 · ξ 1 t 1 s · ξ 11 t , g t , ε t , θ ¯ , ε ˙ t , θ ¯ = 1 , 2 · ξ 1 t 1 s · 0 t , g t , ε t , θ ¯ , ε ˙ t , θ ¯ = 0 , 2 u t θ 1 θ 2 = ξ 2 t 1 s · ξ 1 t G c s , θ ¯ · ξ 12 t , g t , ε t , θ ¯ , ε ˙ t , θ ¯ = 1 , ξ 2 t 1 s · 0 t G c s , θ ¯ · ξ 12 t , g t , ε t , θ ¯ , ε ˙ t , θ ¯ = 0 , 2 u t θ 2 θ 1 = ξ 2 t 1 s · ξ 1 t G c s , θ ¯ · ξ 21 t , g t , ε t , θ ¯ , ε ˙ t , θ ¯ = 1 , ξ 2 t 1 s · 0 t G c s , θ ¯ · ξ 21 t , g t , ε t , θ ¯ , ε ˙ t , θ ¯ = 0 , 2 u t θ 2 θ 2 = 2 s · ξ 2 t G c s , θ ¯ · ξ 22 t , g t , ε t , θ ¯ , ε ˙ t , θ ¯ = 1 , 2 s · 0 t G c s , θ ¯ · ξ 22 t , g t , ε t , θ ¯ , ε ˙ t , θ ¯ = 0 .
The matrix H(J) for model (17) according to criteria (6) is presented similarly to in [25]:
H J = λ · J + 2 · 0 ξ 1 2 t ξ 11 t · ε t , θ ¯ d t λ · J + 2 · 0 ξ 1 t · ξ 2 t ξ 12 t · ε t , θ ¯ d t λ · J + 2 · 0 ξ 1 t · ξ 2 t ξ 21 t · ε t , θ ¯ d t λ · J + 2 · 0 ξ 2 2 t ξ 22 t · ε t , θ ¯ d t
The optimization process is complete if the following conditions are met:
I l I l 1 < ϵ   a n d   n S J N ,
where ϵ is the permissible error and nSJ is the gradient sign change number.
As noted above, in addition to testing the automatic parametric optimization-generated algorithm’s performance and, accordingly, testing the computation (19) correctness, a technique oriented towards a computational experiment was used. The idea behind this technique is that the automatic parametric optimization algorithm is launched with at least three different initial values for the adjustable parameters θ ¯ k 0 = θ 1 k 0 , θ 2 k 0 and k = 1,2 , 3 ¯ ; then, at the adjustable parameters’ space points θ ¯ k * = θ 1 k * , θ 2 k * , the necessary condition for the optimality criterion extremum for θ ¯ * is checked:
J θ ¯ * θ ¯ = 0 ± ,
where Δ is the calculation error.
Additionally, to check the automatic parametric optimization-generated algorithm’s results reliability, the following condition is used:
J θ ¯ * J θ ¯ < δ ,   i , j
for θ ¯ * within the range of adjustable parameters used in this research, where δ is the permissible error.
Based on the above, we summarize that, in this research, the automatic parametric optimization-generated algorithm’s performance indicator comprises criteria (16), (21), and (22) for the fulfilment for θ ¯ * .
In the proposed algorithm, the controller’s parameters are updated according to (10): θ1₎ = θ₎ − ηₗ · (W · ∇J), where the dynamic training rate, ηₗ, is adjusted based on the change in the optimization criterion, which allows for adapting the optimization step as it approaches the minimum of the function J (see (10)–(12)). The stopping conditions specified in (12)–(14) ensure that the iterative process is completed when the difference between the J values at successive steps becomes lower than a given threshold, which prevents overtraining. Regularization is used to stabilize the algorithm: adding the coefficient λ to the Hessian matrix, H(J) in (15), and checking for positive definiteness using Sylvester’s criterion (16) minimizes the influence of noise and nonlinearities. In addition, the choice of switching function, g(t) in (4), where the parameter α determines the exponential decay rate and the coefficients k1, k2, and k3 regulate the gain level, allows one to precisely balance the controller’s speed and stability under dynamically changing control conditions.
To implement the proposed method in neural network form, an architecture is proposed (Figure 3), which includes consideration of adaptive parameters and system nonlinearities, as well as an algorithm for its training (Table 2). The proposed neural network architecture (Figure 3) is built on a multilayer structure, including input, hidden and output layers. The input layer accepts two key parameters: the system error ϵ(t) and the reference action r(t), which the system’s dynamic behaviour determines. The hidden layers consist of three fully connected levels with nonlinear activation functions, such as SmoothReLU and tanh [32,33], to model complex nonlinear relations in the system.
The first hidden layer processes the error ϵ(t) using a nonlinear activation function, simulating fnonline(x, u). The second hidden layer computes the control effect, u(t) = Gc(s) ⋅ ϵ(t). The third hidden layer is responsible for integrating the switching parameters (e.g., via the functions ϕ(t)). To ensure network adaptation, parameters such as kp, Ti, k1, k2, k3, and α are trained as neural network weights. They are optimized using the gradient descent method with regularization carried out by calculating gradients based on the sensitivity function, Ξ(t).
The output layer is designed to calculate the output coordinate x(t) = Gp(s) ⋅ u(t), which depends on the control action u(t) and is described by the control object operator Gp(s), taking into account the delay and nonlinearities. To stabilize the training process, a regularization coefficient λ is introduced, which improves the network’s resistance to overtraining and noise. Additionally, a dynamic training rate ηl is used, depending on the change in the optimization function, which accelerates convergence and improves the neural network’s accuracy.

3. Case Study

Based on the numerous research results devoted to helicopter turboshaft engine (TE) monitoring methods, including the research in [32,33,34,35,36,37,38], the TV3-117 engine, which is part of the Mi-8MTV helicopter power plant [39,40], was chosen as the research object for the computational experiment. The main diagnostic parameter recorded by the standard onboard sensor D-2M [32,41] is the gas-generator rotor speed, nTC. The nTC values were obtained exclusively from the Mi-8MTV helicopter flight test data. Data were collected for 256 s under actual flight conditions at the nominal engine operating mode with a sampling interval of 1 s. The base frequency value, nTCreq, was set equal to 15,900 rpm [31,40]. The data were provided to the authors’ team upon an official request to the Ministry of Internal Affairs of Ukraine within the framework of the project “Theoretical and Applied Aspects of Aviation Sphere Development”, registered under number 0123U104884. The changes in the nTC frequency recorded under the operating conditions of the TV3-117 engine demonstrate the time series complexity (Figure 4), and the constructed diagrams emphasize the importance of the current parameter values and their accumulation in the model’s memory [42,43]. In accordance with the recommendations provided in [32,33,34,35,36,37,38], the 256 nTC values presented in Figure 4 were selected for analysis. The time sampling and adaptive quantization with a dynamic range applied to the reconstructed nTC signal [33] resulted in its discretized form (points in Figure 4).
The generated diagrams (Figure 4) play an essential role in the formation of the training dataset. According to the Nyquist theorem, the sampling frequency should be at least twice the maximum frequency to reconstruct the signal accurately. With 256 measurements over 256 s, the sampling frequency is 1 measurement per second, which is optimal for tracking slowly changing parameters and provides sufficient detail without an excessive amount of data [32,33,34,35,36,37,38]. Thus, the choice of 256 measurements is a compromise between the sampling frequency, data amount, and computing resources, which allows for achieving a balance between data processing accuracy and efficiency.
To form a training dataset consisting of the system error, ϵ(t), and the reference action, r(t), values, z-normalization of the nTC values was carried out [44], with the help of which each nTC value was transformed in such a way that it was expressed in standard deviation units from the mean. This allows us to compile the nTC values into a form with a mean value of 0 and a standard deviation of 1, which is essential for improving the developed neural network training algorithm’s (Table 2) convergence and for correct comparison of the data with different scales. The nTC values’ z-normalization was carried out according to the following expression:
z n T C i = n T C i 1 N · i = 1 N n T C i 1 N · i = 1 N n T C i 1 N · i = 1 N n T C i 2 ,
where N = 256.
To determine the system error, ϵ(t), for nTC values normalized using z-normalization, the mean standard error was applied, calculated as follows:
ϵ i = σ z N ,
where σ z = 1 N · i = 1 N z n T C i z n T C ¯ 2 is the standard deviation of the z-normalized nTC values, and z n T C ¯ is the mean of the z-normalized nTC values, typically z n T C ¯ = 0 if the data are normalized correctly.
The reference action is the nTC value in the helicopter TE’s current operating mode [42]. For example, in the emergency mode, nTC = 97.4% or nTC = 15,486 rpm; in the takeoff mode, nTC = 96.3% or nTC = 15,312 rpm; in the nominal mode, nTC = 94.7% or nTC = 15,057 rpm; in the first cruising mode, nTC = 93.6% or nTC = 14,882 rpm; and in the second cruising mode, nTC = 91.7% or nTC = 14,580 rpm. In this case, a deviation from the specified value of ±0.5% is allowed.
Since the Mi-8MTV helicopter’s flight tests were conducted in the nominal engine operating mode, the reference action is 256 values of nTC = 15,057 rpm ±0.5%, where the deviation of ±0.5% was selected randomly for each i-th nTC value. The nTC values, as a reference action, were also normalized according to (23) and (24). Thus, the training dataset presented in Table 3 is obtained.
Based on [32,33,34,35,36,37,38], the training dataset’s (Table 3) homogeneity was assessed using the Fisher–Pearson [45,46,47] and Fisher–Snedecor [48,49,50] tests at a significance level of α = 0.01 (Table 4). The significance level of 0.01 was chosen to ensure a statistically valid conclusion about high reliability, which is especially important for helicopter flight safety. This strict criterion minimizes the first type of error probability, which contributes to the helicopter TE’s more accurate and safe control under operational conditions [51,52].
The training and test datasets’ representativeness was assessed using cluster analysis (k-means method) [53], which involved dividing the training dataset (Table 3) into k groups (clusters), where each value belongs to the cluster with the highest mean value [49,50]. The algorithm repeats the point distribution, minimizing the distance between them and their centroids until the clusters stabilize [54,55]. The cluster analysis of the training dataset (Table 3) identified eight clusters (I … VIII). The training and test datasets were randomly generated in a 2:1 ratio (67 and 33%, respectively), and the same 8 clusters were found in both sets, indicating their structural similarity. The distances between clusters in the training and test datasets were almost identical, confirming their similarity (Figure 5).
Thus, the optimal size of the training dataset is 256 elements (100%), that of the control dataset is 172 elements (67% of the training dataset size), and that of the test dataset is 84 elements (33% of the training dataset size).
The conducted performance test results include the computation of the H(J) matrix and the transient process diagrams at the automatic parametric optimization algorithm’s initial and final points (Figure 6, Figure 7, Figure 8 and Figure 9) with the above-described reference action. The research object’s (the TV3-117 engine) parameters are as follows: T o b 1 = 1.0, T o b 2 = 4.0, kob = 1.0, and τob = 5.0 [32,33,34,35,36,37,38].
Figure 6 demonstrates that the automatic parametric optimization-developed algorithm of the system with controller (5) ensures θ ¯ * computation from the initial values θ ¯ 0 , at which point the transient process in the automatic system has both oscillatory and monotonic characteristics. Table 5 shows that the matrix H(J) is positive-definite at the algorithm’s endpoints θ ¯ * . Figure 6, Figure 7, Figure 8 and Figure 9 illustrate the algorithm’s (10) convergence patterns. Figure 6 shows that the proposed method’s application allowed the transient process time for the parameter nTC to be reduced by 2 times, while the nTC values were set at the level of ≈1 (i.e., nTC values were restored to the initial ones). The search-free method’s [25,26,27,28] application also allows for a reduction in the transient process time for parameter nTC by 2 times. However, the nTC values were set at the level of ≈ 0.5 (i.e., nTC values were restored to half of the initial ones). The nTC = 0.5 · nTC values used in the real helicopter TE’s operation conditions are impossible.
Figure 7 illustrates the change in the two components of the gradient of the optimization function J during the automatic parametric optimization algorithm’s operation. Figure 7 displays the changes in trajectories in the corresponding gradient components for each of the initial parameters (“set 1” is the black curve, “set 2” is the blue curve, and “set 3” is the green curve) in three sets. Figure 7 shows that, as the iterations progress, the gradient component values gradually decrease, indicating that the algorithm converges to the optimization criterion’s minimum point. The comparison of the curves for parameters with different sets (indicated in Table 5) allows us to estimate the algorithm’s convergence sensitivity to the choice of initial values.
Figure 8 shows the changes’ trajectories in the controller’s adjustable parameter values (e.g., parameters θ1, θ2) during the iterative optimization process. Each curve (black, blue, and green) corresponds to one of the three sets of initial parameters, which allows us to trace how the parameters are gradually adjusted and approach the optimal values, ensuring an improvement in the system’s characteristics. According to Figure 8, the parameter’s observed convergence confirms that the selected update method (based on gradient descent with a dynamic training coefficient) effectively tunes the controller.
Figure 9 demonstrates how the value of the optimization criterion J changes as the algorithm’s iterations are performed. Figure 9 shows that the J value decreases with each iteration and converges to a minimum in about 20 iterations, indicating the algorithm’s rapid convergence. In comparison, in the “search-free” method [25,26,27,28], the criterion’s optimization is achieved in 24 iterations, which emphasizes the proposed method’s advantage (the number of iterations reducing by about 1.2 times). This result confirms the effectiveness of the approach used: a rapid decrease in the J value demonstrates that the algorithm successfully minimizes the optimization function, thereby improving the control characteristics.
According to Table 5, the matrix H(J) is positively defined, which confirms, together with the technique given in [25], that the optimal parameters for controller (5) were found by the automatic parametric optimization-generated algorithm, based on the minimum value of criterion (6).
The presented stability analysis method was carried out under conditions where the object parameters take pre-specified values: T o b 1 = 1.0, T o b 2 = 4.0, kob = 1.0, and τob = 5.0.
The controlled object Gp(p) operator takes into account the system’s inertia, delay and possible nonlinearities. For these parameters, the object is characterized by a significant delay ( τ o b T o b m > 1 ), which increases the control and raises the complexity of the requirements for the controller’s algorithm. The modified PI controller with the switching function g t , θ ¯ (4) is adaptive and allows for taking into account the transient process’s features. In the stability analysis, the choice of parameters k1, k2, k3, and α plays an important role since they determine the controller’s response to error deviations and its change rate. The first- and second-order sensitivity functions (ξi(t) and ξij(t)) show how a change in the controller parameters (θ1, θ2) affects the system’s dynamics. For the given object parameters, the sensitivity function takes into account a significant delay and inertia, which leads to the need to adjust the training rate, ηl, in the gradient procedure (10). Sylvester’s condition (16) is used to check the Hessian matrix’s positive definiteness, which is necessary to confirm that the minimum optimization criterion J was found. For this object, given its parameters, the Hessian matrix determinant and its diagonal elements require additional regularization (λ > 0) (15) to compensate for the significant delay effect. The switching function, g t , θ ¯ , significantly affects the transient processes’ stability. For ε t , θ ¯ · ε ˙ t , θ ¯ 0 , the system demonstrates smoother control due to a decrease in the gain. However, for ε t , θ ¯ · ε ˙ t , θ ¯ < 0 , the presence of exponential damping, eα ·t, can lead to oscillations if the value chosen for the parameter α is too small. The optimization functional J (6) reflects the balance between minimizing the error and control costs. For these parameters of the object, the weighting factors a1 and a2 must be selected in such a way as to take into account the delay’s influence on the control action, u(t).
Based on the above, it was found that the method is stable for the given parameters T o b 1 = 1.0, T o b 2 = 4.0, kob = 1.0, and τob = 5.0, provided that the controller’s parameters (kp, Ti, k1, k2, k3, α) are selected correctly. The significant delay of the object ( τ o b T o b m > 1 ) is levelled out by the use of regularization and an adaptive training rate. The Hessian matrix’s positive definiteness confirms the system’s stability. To further confirm the stability, simulation modelling was carried out using the second-order sensitivity functions according to the developed algorithm (Table 6).
The resulting diagram (Figure 10) for the sensitivity functions demonstrates the system parameters’ influence on its behaviour time dynamics.
The first-order sensitivity function (ξ1(t) and ξ2(t)) diagrams show two curves: the red curve, ξ1(t), is the system’s sensitivity to the first controller parameter, θ1, and the blue curve, ξ2(t), is the system’s sensitivity to the second parameter, θ2. At the beginning of the process (t ≈ 0), the ξ1(t) and ξ2(t) values can be large, indicating the significant influence of parameters θ1 and θ2 on the system dynamics at the initial moment in time. As time increases (t → ∞), both functions tend to zero, meaning a decrease in the parameters’ influence on the system’s output in the steady-state mode. The ξ1(t) values are higher than ξ2(t), which may indicate that the parameter θ1 had a more substantial influence on the system’s dynamics, and the time delay and curve amplitudes differ due to each parameter’s different contributions.
The second-order sensitivity function (ξ11(t), ξ12(t), ξ22(t)) diagrams show three curves: The red curve, ξ11(t), is the sensitivity’s second derivative with respect to θ1; it shows how θ1 influences changes over time. The green curve, ξ12(t), is the second-order joint sensitivity between the parameters θ1 and θ2; it reflects the parameter interactions. The blue curve, ξ22(t), is the sensitivity’s second derivative with respect to θ2; it shows how the influence of θ2 changes over time. The second-order functions ξ11(t) and ξ22(t) have high values at the beginning (t ≈ 0) and then quickly decrease, demonstrating a decrease in the parameters’ influence in the steady-state mode. The joint sensitivity, ξ12(t), has a smaller amplitude, which may indicate a weak relationship between the parameters θ1 and θ2. At the same time, the ξ11(t) value is higher than ξ22(t), confirming the greater significance of θ1 compared to θ2. The ξ12(t) values remain low, indicating weak cross-dependence between the parameters.
Thus, the diagrams show that the parameters θ1 and θ2 have the most significant influence on the system’s dynamics at the initial moment in time. At the same time, the parameter θ1 has a more substantial impact on the system’s behaviour than θ2, both in the first-order and second-order sensitivity terms. In turn, the mutual influence of the parameter ξ12(t) is insignificant, which simplifies the optimization issue since the parameters can be selected almost independently. As a result, it is proven that the output coordinate x(t) stabilizes at optimal q*, and the criterion J is minimized at q* = [0.45, 0.12].
When conducting a computational experiment using the developed neural network (Figure 3) and the developed algorithm for its training (Table 2), a maximum accuracy of 0.993 was achieved on both the training and validation datasets (Figure 11a) with 160 training epochs. At the same time, the loss function drops from 2.5 to 0.5% on both the training and validation datasets (Figure 11b) with 160 training epochs.
The obtained data confirm that the proposed neural network (Figure 3), which trains 160 epochs using the developed training algorithm (Table 3), is sufficient to achieve the minimum value of the error function (MSEmin = 0.005). For comparison, similar experiments were performed using standard training algorithms for this neural network (Figure 3), the results of which are presented in Table 7.
The comparative analysis results showed that the developed training algorithm is the most effective, achieving the error function’s minimum value (MSEmin = 0.005) for the minimum number training epochs (160). Compared with traditional and alternative algorithms, the minimum values of MSEmin achieved by them exceed the developed algorithm’s similar indicator by 1.86 to 6.02 times, while the optimal number of training epochs increases by 1.38…2.38 times.
The developed method for neural network implementation using the developed neural network (Figure 3) comparison was made with other well-known neural network architectures—the traditional RBF network [20,56], Jordan neural network [21,57], Elman neural network [22,58], Hopfield neural network [23,59], and Hamming neural network [24,60]—according to traditional quality metrics: accuracy, precision, recall, F1-score, and AUC-ROC, as well as neural network training time (Table 8). These metrics are calculated according to the following expressions [61,62,63]:
A c c u r a c y = 1 N · i = 1 N 1 n T C i = n ^ T C i , P r e c i s i o n = T P T P + F P ,   R e c a l l = T P T P + F N ,   F 1 s c o r e = 2 · P r e c i s i o n · R e c a l l P r e c i s i o n + R e c a l l ,   A U C R O C = 0 1 T P R · F P R 1 t d t .  
According to the comparative analysis data (Table 8), the developed neural network is trained significantly faster and shows improved quality indicators. Thus, the training time is only 1 min 12 s (72 s), which is approximately 2 times faster than that of the traditional RBF network, whose training time reaches 2 min 25 s (145 s) [20,56]. In comparison with the Jordan and Elman architectures, the training time is reduced by approximately 3.5 times (72 s vs. 257 and 252 s, respectively) [21,22,57,58]. Moreover, in terms of metrics such as accuracy and F1-score, the developed network shows an about 18…22% improvement compared to the Hopfield (accuracy 0.837, F1-score 0.833) and Hamming (accuracy 0.811, F1-score 0.819) architectures [23,24,59,60]. These quantitative improvements indicate that the proposed architecture combines high accuracy and significantly reduced training time, making it more effective for real-time applications.

4. Discussion

4.1. Generalization of Results

The developed method for studying processes in ACSs (see Figure 2) is based on the operator approach used for modelling both the control object and the controller. The processes in the automatic control system are described by the system of equations given in (1), which determines the relationship between the control error, ε(t, θ ¯ ), the reference action, r(t), the control action, u(t), the output coordinate, x(t), and the controller, Gc(s, θ ¯ ), and control object, Gp(s), operators. The control object operator, presented in the form of (2), includes parameters that allow for describing the most complex object dynamics, including the static gain, kob; time constants T o b 1 and T o b 2 ; the delay time, τob; and the nonlinear part, fnonline(x, u). A modification of the traditional PI controller is used for the controller, presented in the form of (3). It includes the switching function g t , θ ¯ , which describes the controller’s adaptation to the current operating conditions. This function is regulated by the parameters k1, k2, k3, and α and is determined according to (4). The control action, u(t), is determined by expression (5), which depends on the error ε(t, θ ¯ ), parameters θ1 and θ2, and the switching function state g t , ε t , θ ¯ , ε ˙ t , θ ¯ . The functional expression in (6), including the weighting coefficients a1 and a2, which take into account the control errors and the control action contributions, is selected as the optimization criterion for adjusting the controller’s parameters. Expression (7) is used to calculate the sensitivity to the parameters θ ¯ , and the update in the parameters’ gradient in the optimization process is determined by expression (10). The conditions for stopping the optimization process are formed on the basis of the analysis of changes in the optimality criterion, as indicated in expressions (12) and (13). The process completion is determined by fulfilling the conditions presented in expression (14). To check the criterion’s minimum, the Hessian matrix H(J) is calculated, whose expression is given in (15). The matrix’s positive definiteness condition is estimated using Sylvester’s criterion, presented in expression (16). The developed method differs from the search-free method [25,26,27,28] by the integration of the nonlinear components into the control object operator, the parameterization of the dynamic training rate ηl, the regularization coefficient λ used to stabilize the optimization, and improved stopping criteria taking into account the permissible error and the number of changes in the gradient sign.
The developed method is implemented as a neural network with a specified architecture (see Figure 3), taking into account the system’s dynamic features and its behavioural nonlinearities. The basis of this architecture is a multilayer structure, where the input layer receives the system error ϵ(t) and the reference action r(t) parameters, and the hidden layers with nonlinear activation functions (SmoothReLU, tanh) model complex relations. A unique feature is the first hidden layer used for processing ϵ(t) with the function fnonline(x, u), the second layer for calculating the regulatory action u(t) = Gc(s) ⋅ ϵ(t), and the third layer for integrating the switching parameters through the functions ϕ(t). Adaptive network parameters such as kp, Ti, k1, k2, k3, and α are trained using the gradient descent method with regularization, where the coefficient λ stabilizes the training, preventing overfitting and noise. The dynamic training rate ηl, depending on the change in the optimization functional, ensures convergence acceleration and increased model accuracy, and the output layer calculates the coordinate x(t), described by the control object operator Gp(s), which allows for taking into account delays and system nonlinearities.
A neural network training algorithm (see Table 2) is developed that includes a multilayer architecture with adaptive parameters and takes into account the system’s nonlinear dynamic features. In the first stage, the network architecture and its parameters (weights and biases) are initialized using the normal distribution or methods such as the Xavier or He methods. In the forward pass stage, the input parameters ϵ(t) and r(t) are sequentially processed through three hidden layers: the first layer computes the nonlinear error processing using SmoothReLU, the second layer models the control action with the tanh function, and the third layer integrates the switching parameters via SmoothReLU. The output layer forms the coordinate x(t) defined by the control object operator. The error is calculated using a loss function, for example, MSE, and the gradients are computed taking into account the sensitivity function, Ξ(t), reflecting the parameters’ effect on the system. In the backpropagation stage, the gradients are updated for all layers using regularization, λ, and a dynamic learning rate, ηl, depending on the change in the optimization function. Adaptive parameters such as kp, Ti, k1, k2, k3, and α are optimized, taking into account the error changes. The training process is repeated until the specified criteria, such as the minimum error or the maximum number of epochs, are achieved, and the algorithm’s accuracy is evaluated on test data, demonstrating the neural network’s ability to effectively handle nonlinearities and system delays.
To conduct the computational experiment, it was noted that the choice of the TV3-117 engine as the research object was based on its widespread use in the Mi-8MTV helicopter power plant, and the key diagnostic parameter of the gas-generator rotor speed, nTC (see Figure 4), recorded by the standard D-2M sensor, was used to form the training dataset (see Table 3). The use of z-normalization made it possible to standardize the data, improving the neural network training algorithm’s convergence (see Table 4), and cluster analysis (see Figure 5) confirmed the representativeness of the training and test datasets. This approach ensures the statistically substantiated reliability of the dataset and error minimization, which are critically important for improving helicopter operation safety.
This research demonstrated the effectiveness of the proposed method for automatic parametric optimization for the TV3-117 engine control system. When analyzing transient processes (see Figure 6, Figure 7, Figure 8 and Figure 9 and Table 5), it provided a twofold reduction in the nTC parameter transition time to a level of ≈1, which confirms its suitability for actual operating conditions. Unlike the search-free method [25,26,27,28], where nTC is restored only to a level of ≈ 0.5, the developed method provides higher-quality control, which makes it practically applicable in aviation technology.
Sensitivity analysis (ξ1, ξ2, ξ11, ξ22, ξ12) showed (see Figure 10 and Table 6) that the parameters θ1 and θ2 have the most significant influence at the initial time, with θ1 dominating, and their interaction is insignificant, which simplifies the tuning of the controller. The iteration process confirmed the positive definiteness of the Hessian matrix, and the method’s convergence was observed to be faster compared to analogues (1.36-fold reduction in iterations). The application of the developed neural network and training algorithm showed a maximum accuracy of 0.993 at 160 epochs (see Figure 11a), ensuring the error function minimization to MSEmin = 0.005 (see Figure 11b). These results confirm the reliability and stability of the proposed method for complex objects with significant delays and inertia.
The comparative analysis (see Table 7) results of the proposed neural network training algorithm and alternative approaches’ efficiency confirmed its superiority. The developed algorithm achieved the mean square error minimum value (MSEmin = 0.005) for the minimum number of training epochs (160), which is significantly better compared to traditional error backpropagation and other methods (exceeding MSEmin by 1.86 to 6.02 times and increasing the number of training epochs by 1.38 to 2.38 times).
The comparison of neural network architectures showed that the developed network has the highest accuracy (0.993), recall (1.0) and F1-measure (0.995) while taking only 1 min and 12 s to train (see Table 8). Alternative networks, such as RBF, Jordan and Elman, have comparable accuracy but require significantly more time to train (from 2 to 4 min). Hopfield and Hamming networks showed substantially lower accuracy values (0.837 and 0.811, respectively), which limits their use for tasks requiring high accuracy.

4.2. Limitations and Further Research

The developed method’s key limitations are its dependence on the control object’s model parameters (such as the static gain, time constants, and nonlinear components) and precise identification, which can reduce the modelling and control accuracy in the event of errors in these parameters [64,65]. The method also requires careful adjustment of the controller’s adaptation parameters and optimization coefficients, which complicates its implementation in practical conditions [66]. Nonlinear components in the control object model may not be applicable to systems with more complex dynamics, and the dependence on the training data (noise, sensor errors) quality can significantly worsen the results [67,68,69]. High computational complexity, requirements for the neural network architecture, and the need for complex optimization of the stopping criteria also limit the method’s applicability, especially in actual conditions with limited computing resources [70].
In addition, there is an overfitting risk when using small or nonuniform datasets, and modelling systems with considerable delays may be inaccurate. The method showed high efficiency for a specific engine type (TV3-117), but its generalizability to other kinds of control systems requires additional adaptation. The algorithm’s implementation in real systems requires powerful computing resources [71,72,73], which may be problematic for existing systems with limited power.
A roadmap for future research to address the developed method’s limitations is presented in Table 9.
The roadmap’s expected results include improvements in the method’s accuracy and reliability, which will allow for its applications to be expanded in various fields [84,85,86], as well as increased computational efficiency and reduced dependence on hardware resources. In addition, we plan to expand the method’s scope, including its adaptation for working with complex dynamic systems with long delays [87,88,89], which will increase the universality and practical value of the approach.

5. Conclusions

A method for researching processes in automatic control systems using an operator approach and a modification of the PI controller has been developed, which allows for the control and regulation of the efficient modelling of object dynamics. The control object operator includes parameters such as static gain and time constants (e.g., T o b 1 = 0.5, T o b 2 = 0.3) and a nonlinear part fnonline(x, u), which makes it possible to take into account various aspects of the system. The PI controller modification, which includes a switching function, adapts the system to changing conditions, increasing the regulation flexibility and accuracy. The algorithm for optimizing the controller’s parameters using gradient descent and regularization (λ = 0.01) stabilizes the training process, minimizing overtraining and data errors.
The developed method’s application to the TV3-117 engine control system as an example demonstrated high efficiency. The algorithm significantly improved transient processes, reducing the gas-generator rotor speed parameter transition time from 2.0 to 1.0 and increasing the control accuracy. At the same time, the neural network, with the use of adaptive parameters (kp = 1.2, Ti = 0.8, k1 = 0.5, k2 = 0.3, k3 = 0.7, α = 0.2) and a special architecture, ensured the error function’s minimization to the value MSEmin = 0.005 for 160 epochs, which confirms the method’s high efficiency and stability for the current operating conditions. Comparative analysis showed that the proposed method outperforms alternative approaches in terms of accuracy (accuracy = 0.993), convergence rate (reduction in the number of epochs from 380 to 160) and the number of training epochs, which makes it promising for use in aviation and other complex dynamic systems.
The developed method has a number of key limitations, including high dependence on the precise identification of the controlled object’s model parameters, such as static gain, time constants and nonlinear components, which can reduce the accuracy with errors in these parameters. The method also requires careful tuning of the controller’s adaptation parameters and optimization coefficients, which complicates its implementation in practical conditions. High computational complexity and requirements for the neural network architecture limit its applicability, especially under limited computing resource conditions. Despite its effectiveness for the TV3-117 engine control system, the method has risks of overfitting with small or heterogeneous datasets, as well as modelling accuracy problems for systems with significant delays and complex dynamics, which requires additional adaptation for other types of control systems and implementation in real systems with limited power.
Prospects for further research include the further development of the method to improve the model’s accuracy and robustness to errors in the control object parameters, as well as simplifying the controller’s adaptation process using methods for automatic parameter adjustment. Critical areas include improving the processing of complex nonlinearities in the object dynamics using more complicated neural network architectures and increasing the method’s robustness to noise and data errors. We also plan to optimize the computational processes to reduce computational costs and improve the method’s generalizability for applications to various types of systems with significant delays and limited computing resources.

Author Contributions

Conceptualization, S.V.; methodology, S.V.; software, L.S., N.S.-Ś., A.S. and V.V.; validation, L.S., N.S.-Ś. and A.S.; formal analysis, S.V.; investigation, L.S. and N.S.-Ś.; resources, L.S., N.S.-Ś. and A.S.; data curation, S.V., N.S.-Ś., A.S. and V.V.; writing—original draft preparation, S.V.; writing—review and editing, L.S., N.S.-Ś., A.S. and V.V.; visualization, L.S., N.S.-Ś., A.S. and V.V.; supervision, L.S., N.S.-Ś., A.S. and V.V.; project administration, S.V.; funding acquisition, L.S. and N.S.-Ś. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Acknowledgments

The Ministry of Internal Affairs of Ukraine supported the research entitled “Theoretical and applied aspects of the development of the aviation sphere” under Project No. 0123U104884. Also, the authors would like to thank the reviewers for their precise and concise recommendations that improved the presentation of the results obtained.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Kulyk, M.S.; Kozlov, V.V.; Volianska, L.G. Automation control system of technical condition of gas turbine engine compressor. Aerosp. Tech. Technol. 2019, 8, 121–128. [Google Scholar] [CrossRef]
  2. Tovkach, S.S. Control laws of the aviation gas turbine engine. Electron. Control. Syst. 2022, 2, 20–25. [Google Scholar] [CrossRef]
  3. Chao, C.-Y.; Chen, C.-W.; Wang, B.-C. Variable Structure Control of a Turbojet Engine. In Proceedings of the 1986 American Control Conference, Seattle, WA, USA, 18–20 June 1986; pp. 482–487. [Google Scholar] [CrossRef]
  4. Delgado-Reyes, G.; Guevara-Lopez, P.; Loboda, I.; Hernandez-Gonzalez, L.; Ramirez-Hernandez, J.; Valdez-Martinez, J.-S.; Lopez-Chau, A. State Vector Identification of Hybrid Model of a Gas Turbine by Real-Time Kalman Filter. Mathematics 2020, 8, 659. [Google Scholar] [CrossRef]
  5. Nadweh, S.; Khaddam, O.; Hayek, G.; Atieh, B.; Haes Alhelou, H. Optimization of P & PI Controller Parameters for Variable Speed Drive Systems Using a Flower Pollination Algorithm. Heliyon 2020, 6, e04648. [Google Scholar] [CrossRef]
  6. Mešanović, A.; Münz, U.; Szabo, A.; Mangold, M.; Bamberger, J.; Metzger, M.; Heyde, C.; Krebs, R.; Findeisen, R. Structured Controller Parameter Tuning for Power Systems. Control Eng. Pract. 2020, 101, 104490. [Google Scholar] [CrossRef]
  7. Sun, J.; Liu, J.; Miao, M.; Lin, H. Research on Parameter Optimization Method of Sliding Mode Controller for the Grid-Connected Composite Device Based on IMFO Algorithm. Sensors 2022, 23, 149. [Google Scholar] [CrossRef]
  8. Kulikov, V.V.; Kutsyi, A.P.; Kutsyi, N.N. The Gradient-Based Algorithm for Parametric Optimization of a Variable Structure PI Controller with Dead Band. Mekhatronika Avtom. Upr. 2020, 21, 530–534. [Google Scholar] [CrossRef]
  9. Liang, X.; Bao, D.; Ge, S.S. Modeling of Neuro-Fuzzy System with Optimization Algorithm as a Support in System Boundary Capability Online Assessment. IEEE Trans. Circuits Syst. II Express Briefs 2023, 70, 2974–2978. [Google Scholar] [CrossRef]
  10. Kinga, S.; Megahed, T.F.; Kanaya, H.; Mansour, D.-E.A. A New Voltage Sensitivity-Based Distributed Feedback Online Optimization for Voltage Control in Active Distribution Networks. Comput. Electr. Eng. 2024, 119, 109574. [Google Scholar] [CrossRef]
  11. Toan, N.T.; Thuy, L.Q.; Kim, D.S. Sensitivity Analysis in Parametric Multiobjective Discrete-Time Control via Fréchet Subdifferential Calculus of the Frontier Map. J. Comput. Appl. Math. 2023, 418, 114662. [Google Scholar] [CrossRef]
  12. Xu, Y.; Tang, H.; Chen, M. Design Method of Optimal Control Schedule for the Adaptive Cycle Engine Steady-State Performance. Chin. J. Aeronaut. 2022, 35, 148–164. [Google Scholar] [CrossRef]
  13. Vladov, S.; Shmelov, Y.; Yakovliev, R. Optimization of Helicopters Aircraft Engine Working Process Using Neural Networks Technologies. CEUR Workshop Proc. 2022, 3171, 1639–1656. Available online: https://ceur-ws.org/Vol-3171/paper117.pdf (accessed on 23 December 2024).
  14. Zhen, M.; Dong, X.; Liu, X.; Tan, C. Accelerated Formulation of Optimal Control Law for Adaptive Cycle Engines: A Novel Design Methodology. Aerosp. Sci. Technol. 2024, 148, 109076. [Google Scholar] [CrossRef]
  15. Nikolaidis, T.; Li, Z.; Jafari, S. Advanced Constraints Management Strategy for Real-Time Optimization of Gas Turbine Engine Transient Performance. Appl. Sci. 2019, 9, 5333. [Google Scholar] [CrossRef]
  16. Cannarsa, P.; Frankowska, H.; Scarinci, T. Second-Order Sensitivity Relations and Regularity of the Value Function for Mayer’s Problem in Optimal Control. SIAM J. Control Optim. 2015, 53, 3642–3672. [Google Scholar] [CrossRef]
  17. Deng, L. Second Order Necessary Conditions and Sensitivity Relations for Optimal Control Problems on Riemannian Manifolds. In Proceedings of the 2018 37th Chinese Control Conference (CCC), Wuhan, China, 25–27 July 2018; pp. 2138–2143. [Google Scholar] [CrossRef]
  18. Vassiliadis, V.S.; Canto, E.B.; Banga, J.R. Second-Order Sensitivities of General Dynamic Systems with Application to Optimal Control Problems. Chem. Eng. Sci. 1999, 54, 3851–3860. [Google Scholar] [CrossRef]
  19. Liang, X.; Bao, D.; Yang, Z. State Evaluation Method for Complex Task Network Models. Inf. Sci. 2024, 653, 119796. [Google Scholar] [CrossRef]
  20. Slema, S.; Errachdi, A.; Benrejeb, M. A Radial Basis Function Neural Network Model Reference Adaptive Controller for Nonlinear Systems. In Proceedings of the 2018 15th International Multi-Conference on Systems, Signals & Devices (SSD), Yasmine Hammamet, Tunisia, 19–22 March 2018. [Google Scholar] [CrossRef]
  21. Zarzycki, K.; Ławryńczuk, M. Long Short-Term Memory Neural Networks for Modeling Dynamical Processes and Predictive Control: A Hybrid Physics-Informed Approach. Sensors 2023, 23, 8898. [Google Scholar] [CrossRef]
  22. Fetanat, M.; Stevens, M.; Jain, P.; Hayward, C.; Meijering, E.; Lovell, N.H. Fully Elman Neural Network: A Novel Deep Recurrent Neural Network Optimized by an Improved Harris Hawks Algorithm for Classification of Pulmonary Arterial Wedge Pressure. IEEE Trans. Biomed. Eng. 2022, 69, 1733–1744. [Google Scholar] [CrossRef]
  23. Atencia, M.; Joya, G. Hopfield networks: From optimization to adaptive control. In Proceedings of the 2015 International Joint Conference on Neural Networks (IJCNN), Killarney, Ireland, 12–17 July 2015. [Google Scholar] [CrossRef]
  24. Lee, G.K. On the use of Hamming distance tuning for the generalized adaptive neural network fuzzy inference controller with evolutionary simulated annealing. In Proceedings of the 2011 IEEE International Conference on Information Reuse & Integration, Las Vegas, NV, USA, 3–5 August 2011. [Google Scholar] [CrossRef]
  25. Kulikov, V.V.; Kutsyi, N.N. Search-free algorithm for parametric optimization of pi controller with semi-permanent integration. Ipolytech J. 2018, 22, 98–108. [Google Scholar] [CrossRef]
  26. Kutsyi, N.N.; Osipova, E.A. The sensitivity analyzers of cascade control system with two integral pulse-duration controllers of cable insulation thickness stabilization. Modern Technologies. System Analysis. Modeling 2011, 4, 111–117. [Google Scholar]
  27. Kulikov, V.V.; Kutsyi, N.N. Application of the extended frequency response method for parametric synthesis of a proportional-integral difference controller. Inf. Math. Technol. Sci. Manag. 2021, 1, 37–42. [Google Scholar]
  28. Kulikov, V.V.; Kutsyi, N.N.; Osipova, E.A. Parametric Optimization of the PID Controller with Restriction Based on the Method of Conjugate Polak–Polyak–Ribier Gradients. Mekhatronika Avtom. Upr. 2023, 24, 240–248. [Google Scholar] [CrossRef]
  29. Kuz’michev, V.; Krupenich, I.; Filinov, E.; Tkachenko, A. Optimization of Gas Turbine Engine Control Using Dynamic Programming. MATEC Web Conf. 2018, 220, 03002. [Google Scholar] [CrossRef]
  30. Kulikov, V.V.; Kutsyi, A.P.; Kutsyi, N.N. Formation of Algorithm of Automatic Parametric Optimization of PI Controller with Variable Parameters While Using Internal Model Control. IOP Conf. Ser. Mater. Sci. Eng. 2021, 1151, 012031. [Google Scholar] [CrossRef]
  31. Kulikov, V.V.; Kutsyi, N.N.; Podkorytov, A.A. Gradient-Based Algorithm for Parametric Optimization of Variable-Structure PI Controller When Using a Reference Model. Adv. Intell. Syst. Comput. 2020, 1295, 938–949. [Google Scholar] [CrossRef]
  32. Vladov, S.; Scislo, L.; Sokurenko, V.; Muzychuk, O.; Vysotska, V.; Osadchy, S.; Sachenko, A. Neural Network Signal Integration from Thermogas-Dynamic Parameter Sensors for Helicopters Turboshaft Engines at Flight Operation Conditions. Sensors 2024, 24, 4246. [Google Scholar] [CrossRef]
  33. Vladov, S.; Scislo, L.; Sokurenko, V.; Muzychuk, O.; Vysotska, V.; Sachenko, A.; Yurko, A. Helicopter Turboshaft Engines’ Gas Generator Rotor R.P.M. Neuro-Fuzzy On-Board Controller Development. Energies 2024, 17, 4033. [Google Scholar] [CrossRef]
  34. Vladov, S.; Sachenko, A.; Sokurenko, V.; Muzychuk, O.; Vysotska, V. Helicopters Turboshaft Engines Neural Network Modeling under Sensor Failure. J. Sens. Actuator Netw. 2024, 13, 66. [Google Scholar] [CrossRef]
  35. Vladov, S.; Shmelov, Y.; Yakovliev, R.; Stushchankyi, Y.; Havryliuk, Y. Neural Network Method for Controlling the Helicopters Turboshaft Engines Free Turbine Speed at Flight Modes. CEUR Workshop Proc. 2023, 3426, 89–108. Available online: https://ceur-ws.org/Vol-3426/paper8.pdf (accessed on 30 December 2024).
  36. Vladov, S.; Shmelov, Y.; Petchenko, M. A Neuro-Fuzzy Expert System for the Control and Diagnostics of Helicopters Aircraft Engines Technical State. CEUR Workshop Proc. 2021, 3013, 40–52. Available online: https://ceur-ws.org/Vol-3013/20210040.pdf (accessed on 2 January 2025).
  37. Vladov, S.; Yakovliev, R.; Hubachov, O.; Rud, J. Neuro-Fuzzy System for Detection Fuel Consumption of Helicopters Turboshaft Engines. CEUR Workshop Proc. 2024, 3628, 55–72. Available online: https://ceur-ws.org/Vol-3628/paper5.pdf (accessed on 5 January 2025).
  38. Vladov, S.; Petchenko, M.; Shmelov, Y.; Drozdova, S.; Yakovliev, R. Helicopters Turboshaft Engines Parameters Identification at Flight Modes Using Neural Networks. In Proceedings of the 2022 IEEE 17th International Conference on Computer Sciences and Information Technologies (CSIT), Lviv, Ukraine, 10–12 November 2022; pp. 5–8. [Google Scholar] [CrossRef]
  39. Pasieka, M.; Grzesik, N.; Kuźma, K. Simulation modeling of fuzzy logic controller for aircraft engines. Int. J. Comput. 2017, 16, 27–33. [Google Scholar] [CrossRef]
  40. Aygun, H.; Caliskan, H. Evaluating and Modelling of Thermodynamic and Environmental Parameters of a Gas Turbine Engine and Its Components. J. Clean. Prod. 2022, 365, 132762. [Google Scholar] [CrossRef]
  41. Vlasenko, D.; Inkarbaieva, O.; Peretiatko, M.; Kovalchuk, D.; Sereda, O. Helicopter Radio System for Low Altitudes and Flight Speed Measuring with Pulsed Ultra-Wideband Stochastic Sounding Signals and Artificial Intelligence Elements. Radioelectron. Comput. Syst. 2023, 3, 48–59. [Google Scholar] [CrossRef]
  42. Vladov, S.; Banasik, A.; Sachenko, A.; Kempa, W.M.; Sokurenko, V.; Muzychuk, O.; Pikiewicz, P.; Molga, A.; Vysotska, V. Intelligent Method of Identifying the Nonlinear Dynamic Model for Helicopter Turboshaft Engines. Sensors 2024, 24, 6488. [Google Scholar] [CrossRef]
  43. Kovtun, V.; Grochla, K.; Połys, K. Investigation of the Information Interaction of the Sensor Network End IoT Device and the Hub at the Transport Protocol Level. Electronics 2023, 12, 4662. [Google Scholar] [CrossRef]
  44. Rusyn, B.; Lutsyk, O.; Kosarevych, R.; Kapshii, O.; Karpin, O.; Maksymyuk, T.; Gazda, J. Rethinking Deep CNN Training: A Novel Approach for Quality-Aware Dataset Optimization. IEEE Access 2024, 12, 137427–137438. [Google Scholar] [CrossRef]
  45. Balakrishnan, N.; Voinov, V.; Nikulin, M.S. Chapter 2—Pearson’s Sum and Pearson-Fisher Test. In Chi-Squared Goodness of Fit Tests with Applications; Balakrishnan, N., Voinov, V., Nikulin, M.S., Eds.; Academic Press: Waltham, MA, USA, 2013; pp. 11–26. [Google Scholar] [CrossRef]
  46. Kim, H.-Y. Statistical Notes for Clinical Researchers: Chi-Squared Test and Fisher’s Exact Test. Restor. Dent. Endod. 2017, 42, 152. [Google Scholar] [CrossRef]
  47. Dyvak, M.; Pukas, A.; Oliynyk, I.; Melnyk, A. Selection the “saturated” block from interval system of linear algebraic equations for recurrent laryngeal nerve identification. In Proceedings of the 2018 IEEE Second International Conference on Data Stream Mining & Processing (DSMP), Lviv, Ukraine; 2018; pp. 444–448. [Google Scholar] [CrossRef]
  48. Berko, A.; Alieksieiev, V.; Holdovanskyi, V. Determination-based correlation coefficient. CEUR Workshop Proc. 2024, 3711, 198–224. Available online: https://ceur-ws.org/Vol-3711/paper12.pdf (accessed on 11 January 2025).
  49. Morozov, V.V.; Kalnichenko, O.V.; Mezentseva, O.O. The method of interaction modeling on basis of deep learning the neural networks in complex IT-projects. Int. J. Comput. 2020, 19, 88–96. [Google Scholar] [CrossRef]
  50. Stefanovic, C.M.; Armada, A.G.; Costa-Perez, X. Second Order Statistics of -Fisher-Snedecor Distribution and Their Application to Burst Error Rate Analysis of Multi-Hop Communications. IEEE Open J. Commun. Soc. 2022, 3, 2407–2424. [Google Scholar] [CrossRef]
  51. de Voogt, A.; Nero, K. Technical Failures in Helicopters: Non-Powerplant-Related Accidents. Safety 2023, 9, 10. [Google Scholar] [CrossRef]
  52. de Voogt, A.; Amour, E.S. Safety of Twin-Engine Helicopters: Risks and Operational Specificity. Saf. Sci. 2021, 136, 105169. [Google Scholar] [CrossRef]
  53. Rusyn, B.; Lutsyk, O.; Kosarevych, R.; Obukh, Y. Application Peculiarities of Deep Learning Methods in the Problem of Big Datasets Classification. Lect. Notes Electr. Eng. 2021, 831, 493–506. [Google Scholar] [CrossRef]
  54. Lytvyn, V.; Dudyk, D.; Peleshchak, I.; Peleshchak, R.; Pukach, P. Influence of the Number of Neighbours on the Clustering Metric by Oscillatory Chaotic Neural Network with Dipole Synaptic Connections. CEUR Workshop Proc. 2024, 3664, 24–34. Available online: https://ceur-ws.org/Vol-3664/paper3.pdf (accessed on 13 January 2025).
  55. Babichev, S.; Krejci, J.; Bicanek, J.; Lytvynenko, V. Gene expression sequences clustering based on the internal and external clustering quality criteria. In Proceedings of the 2017 12th International Scientific and Technical Conference on Computer Sciences and Information Technologies (CSIT), Lviv, Ukraine, 5–8 September 2017. [Google Scholar] [CrossRef]
  56. Wong, H.-T.; Mai, J.; Wang, Z.; Leung, C.-S. Generalized M-Sparse Algorithms for Constructing Fault Tolerant RBF Networks. Neural Netw. 2024, 180, 106633. [Google Scholar] [CrossRef]
  57. Wysocki, A.; Lawrynczuk, M. Jordan Neural Network for Modelling and Predictive Control of Dynamic Systems. In Proceedings of the 2015 20th International Conference on Methods and Models in Automation and Robotics (MMAR), Miedzyzdroje, Poland, 24–27 August 2015; pp. 145–150. [Google Scholar] [CrossRef]
  58. Ren, G.; Cao, Y.; Wen, S.; Huang, T.; Zeng, Z. A Modified Elman Neural Network with a New Learning Rate Scheme. Neurocomputing 2018, 286, 11–18. [Google Scholar] [CrossRef]
  59. Checiu, D.; Bode, M.; Khalil, R. Reconstructing Creative Thoughts: Hopfield Neural Networks. Neurocomputing 2024, 575, 127324. [Google Scholar] [CrossRef]
  60. Khristodulo, O.I.; Makhmutov, A.A.; Sazonova, T.V. Use Algorithm Based at Hamming Neural Network Method for Natural Objects Classification. Procedia Comput. Sci. 2017, 103, 388–395. [Google Scholar] [CrossRef]
  61. Turchenko, V.; Chalmers, E.; Luczak, A. A deep convolutional auto-encoder with pooling—Unpooling layers in caffe. Int. J. Comput. 2019, 1, 8–31. [Google Scholar] [CrossRef]
  62. Kovtun, V.; Altameem, T.; Al-Maitah, M.; Kempa, W. Entropy-Metric Estimation of the Small Data Models with Stochastic Parameters. Heliyon 2024, 10, e24708. [Google Scholar] [CrossRef] [PubMed]
  63. Cherrat, E.M.; Alaoui, R.; Bouzahir, H. Score fusion of finger vein and face for human recognition based on convolutional neural network model. Int. J. Comput. 2020, 19, 11–19. [Google Scholar] [CrossRef]
  64. Kosarevych, R.; Lutsyk, O.; Rusyn, B.; Alokhina, O.; Maksymyuk, T.; Gazda, J. Spatial Point Patterns Generation on Remote Sensing Data Using Convolutional Neural Networks with Further Statistical Analysis. Sci. Rep. 2022, 12, 14341. [Google Scholar] [CrossRef]
  65. Sholomii, Y.; Yakovyna, V. Quality Assessment and Assurance of Machine Learning Systems: A Comprehensive Approach. Commun. Comput. Inf. Sci. 2023, 1980, 265–275. [Google Scholar] [CrossRef]
  66. Burov, Y. The Introduction of Attentional Mechanism in the Situational Awareness Process. CEUR Workshop Proc. 2022, 3171, 1076–1086. Available online: https://ceur-ws.org/Vol-3171/paper78.pdf (accessed on 15 January 2025).
  67. Berko, A.; Alieksieiev, V.; Dovbysh, A. Performance evaluation and analysis with code benchmarking and generative AI. CEUR Workshop Proc. 2024, 3711, 169–183. Available online: https://ceur-ws.org/Vol-3711/paper10.pdf (accessed on 15 January 2025).
  68. Bashtyk, Y.; Campos, J.; Fechan, A.; Konstantyniv, S.; Yakovyna, V. Computer monitoring of physical and chemical parameters of the environment using computer vision systems: Problems and prospects. CEUR Workshop Proc. 2020, 2753, 437–442. [Google Scholar]
  69. Shakhovska, N.; Yakovyna, V. Feature Selection and Software Defect Prediction by Different Ensemble Classifiers. Lect. Notes Comput. Sci. 2021, 12923, 307–313. [Google Scholar] [CrossRef]
  70. Shakhovska, N.; Yakovyna, V.; Kryvinska, N. An improved software defect prediction algorithm using self-organizing maps combined with hierarchical clustering and data preprocessing. Lect. Notes Comput. Sci. 2020, 12391, 414–424. [Google Scholar] [CrossRef]
  71. Lipyanina, H.; Sachenko, S.; Lendyuk, T.; Sachenko, T. Targeting Model of HEI Video Marketing based on Classification Tree. CEUR Workshop Proc. 2020, 2732, 487–498. Available online: https://ceur-ws.org/Vol-2732/20200487.pdf (accessed on 16 January 2025).
  72. Heßling, H. The challenge of managing and analyzing big data. Int. J. Comput. 2014, 12, 204–209. [Google Scholar] [CrossRef]
  73. Siroky, P.; Hartansky, R.; Petrilak, J. Possibilities of increasing the reliability by methods of software and time redundancy. Int. J. Comput. 2014, 2, 35–40. [Google Scholar] [CrossRef]
  74. Fan, L.; Yang, G.; Zhang, Y.; Gao, L.; Wu, B. A Novel Tolerance Optimization Approach for Compressor Blades: Incorporating the Measured out-of-Tolerance Error Data and Aerodynamic Performance. Aerosp. Sci. Technol. 2025, 158, 109920. [Google Scholar] [CrossRef]
  75. Dyvak, M.; Manzhula, V.; Melnyk, A.; Rusyn, B.; Spivak, I. Modeling the Efficiency of Biogas Plants by Using an Interval Data Analysis Method. Energies 2024, 17, 3537. [Google Scholar] [CrossRef]
  76. Shubyn, B.; Maksymyuk, T.; Gazda, J.; Rusyn, B.; Mrozek, D. Federated Learning: A Solution for Improving Anomaly Detection Accuracy of Autonomous Guided Vehicles in Smart Manufacturing. Lect. Notes Electr. Eng. 2024, 1198, 746–761. [Google Scholar] [CrossRef]
  77. Bodyanskiy, Y.; Kostiuk, S. Learnable Extended Activation Function for Deep Neural Networks. Int. J. Comput. 2023, 22, 311–318. [Google Scholar] [CrossRef]
  78. Kosarevych, R.; Lutsyk, O.; Rusyn, B. Detection of pixels corrupted by impulse noise using random point patterns. Vis. Comput. 2022, 38, 3719–3730. [Google Scholar] [CrossRef]
  79. Bodyanskiy, Y.; Shafronenko, A.; Pliss, I. Clusterization of Vector and Matrix Data Arrays Using the Combined Evolutionary Method of Fish Schools. Syst. Res. Inf. Technol. 2022, 4, 79–87. [Google Scholar] [CrossRef]
  80. Marakhimov, A.R.; Khudaybergenov, K.K. Approach to the synthesis of neural network structure during classification. Int. J. Comput. 2020, 19, 20–26. [Google Scholar] [CrossRef]
  81. Bodyanskiy, Y.V.; Tyshchenko, O.K. A Hybrid Cascade Neuro–Fuzzy Network with Pools of Extended Neo–Fuzzy Neurons and Its Deep Learning. Int. J. Appl. Math. Comput. Sci. 2019, 29, 477–488. [Google Scholar] [CrossRef]
  82. Vladov, S.; Shmelov, Y.; Yakovliev, R.; Petchenko, M.; Drozdova, S. Neural Network Method for Helicopters Turboshaft Engines Working Process Parameters Identification at Flight Modes. In Proceedings of the 2022 IEEE 4th International Conference on Modern Electrical and Energy System (MEES), Kremenchuk, Ukraine, 20–23 October 2022; pp. 604–609. [Google Scholar] [CrossRef]
  83. Vladov, S.; Shmelov, Y.; Yakovliev, R. Methodology for Control of Helicopters Aircraft Engines Technical State in Flight Modes Using Neural Networks. CEUR Workshop Proc. 2022, 3137, 108–125. [Google Scholar] [CrossRef]
  84. Baranovskyi, D.; Bulakh, M.; Myamlin, S.; Kebal, I. New Design of the Hatch Cover to Increase the Carrying Capacity of the Gondola Car. Adv. Sci. Technol. Res. J. 2022, 16, 186–191. [Google Scholar] [CrossRef]
  85. Sagin, S.; Madey, V.; Sagin, A.; Stoliaryk, T.; Fomin, O.; Kučera, P. Ensuring Reliable and Safe Operation of Trunk Diesel Engines of Marine Transport Vessels. J. Mar. Sci. Eng. 2022, 10, 1373. [Google Scholar] [CrossRef]
  86. Sagin, S.V.; Sagin, S.S.; Fomin, O.; Gaichenia, O.; Zablotskyi, Y.; Píštěk, V.; Kučera, P. Use of Biofuels in Marine Diesel Engines for Sustainable and Safe Maritime Transport. Renew. Energy 2024, 224, 120221. [Google Scholar] [CrossRef]
  87. Nazarkevych, M.; Kowalska-Styczen, A.; Lytvyn, V. Research of Facial Recognition Systems and Criteria for Identification. In Proceedings of the IEEE International Conference on Intelligent Data Acquisition and Advanced Computing Systems: Technology and Applications, IDAACS, Dortmund, Germany, 7–9 September 2023; pp. 555–558. [Google Scholar] [CrossRef]
  88. Kamran, M.; Nadeem, M.; Żywiołek, J.; Abdalla, M.E.M.; Uzair, A.; Ishtiaq, A. Enhancing Transportation Efficiency with Interval-Valued Fermatean Neutrosophic Numbers: A Multi-Item Optimisation Approach. Symmetry 2024, 16, 766. [Google Scholar] [CrossRef]
  89. Denizci, A.; Karadeniz, S.; Ulu, C. Fuzzy Cognitive Map Based PI Controller Design. Adv. Intell. Syst. Comput. 2020, 1197, 1250–1257. [Google Scholar] [CrossRef]
Figure 1. Structural diagram of the researched PI controller with semi-permanent integration.
Figure 1. Structural diagram of the researched PI controller with semi-permanent integration.
Applsci 15 02586 g001
Figure 2. Pseudocode illustrating the steps of gradient-based optimization.
Figure 2. Pseudocode illustrating the steps of gradient-based optimization.
Applsci 15 02586 g002
Figure 3. The proposed neural network architecture.
Figure 3. The proposed neural network architecture.
Applsci 15 02586 g003
Figure 4. The gas-generator rotor speed parameter recorded onboard the helicopter during the 256 s research interval.
Figure 4. The gas-generator rotor speed parameter recorded onboard the helicopter during the 256 s research interval.
Applsci 15 02586 g004
Figure 5. Cluster analysis results: (a) training dataset, (b) test dataset.
Figure 5. Cluster analysis results: (a) training dataset, (b) test dataset.
Applsci 15 02586 g005
Figure 6. The transient processes’ nTC(t) diagrams, reflecting the dynamics in the initial (black curves) and final (blue curves) states of the automatic parametric optimization algorithm’s operation: 1—parameter set 1; 2—parameter set 2; 3—parameter set 3 (note: the values of parameter sets 1, 2, and 3 are presented in Table 5).
Figure 6. The transient processes’ nTC(t) diagrams, reflecting the dynamics in the initial (black curves) and final (blue curves) states of the automatic parametric optimization algorithm’s operation: 1—parameter set 1; 2—parameter set 2; 3—parameter set 3 (note: the values of parameter sets 1, 2, and 3 are presented in Table 5).
Applsci 15 02586 g006
Figure 7. Diagrams of the values of the gradient components, J θ 1 l (a) and J θ 2 l (b) of the optimization criterion J during the automatic parametric optimization algorithm’s operation: (black curve)—parameter set 1; (blue curve)—parameter set 2; (green curve)—parameter set 3 (note: the values of parameter sets 1, 2, and 3 are presented in Table 5).
Figure 7. Diagrams of the values of the gradient components, J θ 1 l (a) and J θ 2 l (b) of the optimization criterion J during the automatic parametric optimization algorithm’s operation: (black curve)—parameter set 1; (blue curve)—parameter set 2; (green curve)—parameter set 3 (note: the values of parameter sets 1, 2, and 3 are presented in Table 5).
Applsci 15 02586 g007
Figure 8. Diagrams of the controller-adjustable parameters θ 1 l (a) and θ 2 l (b) during the automatic parametric optimization algorithm’s operation: (black curve)—parameter set 1; (blue curve)—parameter set 2; (green curve)—parameter set 3 (note: the values of parameter sets 1, 2, and 3 are presented in Table 5).
Figure 8. Diagrams of the controller-adjustable parameters θ 1 l (a) and θ 2 l (b) during the automatic parametric optimization algorithm’s operation: (black curve)—parameter set 1; (blue curve)—parameter set 2; (green curve)—parameter set 3 (note: the values of parameter sets 1, 2, and 3 are presented in Table 5).
Applsci 15 02586 g008
Figure 9. Diagrams of the optimization criterion values during the automatic parametric optimization algorithm’s operation: (black curve)—parameter set 1; (blue curve)—parameter set 2; (green curve)—parameter set 3 (note: the values of parameter sets 1, 2, and 3 are presented in Table 5).
Figure 9. Diagrams of the optimization criterion values during the automatic parametric optimization algorithm’s operation: (black curve)—parameter set 1; (blue curve)—parameter set 2; (green curve)—parameter set 3 (note: the values of parameter sets 1, 2, and 3 are presented in Table 5).
Applsci 15 02586 g009
Figure 10. Schemes follow the same formatting.
Figure 10. Schemes follow the same formatting.
Applsci 15 02586 g010aApplsci 15 02586 g010b
Figure 11. Diagram of the quality metrics: (a) accuracy metric diagram, (b) loss metric diagram.
Figure 11. Diagram of the quality metrics: (a) accuracy metric diagram, (b) loss metric diagram.
Applsci 15 02586 g011
Table 1. The developed algorithm for the proposed neural network training.
Table 1. The developed algorithm for the proposed neural network training.
ApproachAdvantagesDisadvantagesStudies
Gradient optimization methodsHigh accuracy in calculating the gradient quality functional in ACSs using first-order sensitivity.Limited applications in systems with pronounced nonlinearity, high computational complexity in multidimensional systems.[14,15]
Methods of sensitivity theoryTakes into account parameter interactions and their influence on changes in the system by calculating second-order sensitivity functions.High computational load, implementation complexity, need for specialized software.[16,17,18]
Optimal control methodsEffective for stable control objects; they take into account control errors and the control signal impact.Scalability issues, difficulties in integration with real-time adaptive controllers.[10,11,12,13]
Neural networks with fixed switching functionsAbility to process large amounts of data and change parameters in real time.Limited ability to take into account nonlinearities and changes in the dynamics of the control object.[19,20,21,22,23,24]
Controllers with variable structure and semi-permanent integrationFlexible adaptation of system parameters depending on the current state; improved tuning accuracy.Complexity in setting up switching functions, high computational complexity in analytical calculation of second-order sensitivity functions.[3,4]
Table 2. The developed algorithm for the proposed neural network training.
Table 2. The developed algorithm for the proposed neural network training.
Step NumberStep NameDescription
1Network initializationThe network architecture consists of an input layer (x), three hidden layers (h1, h2, h3), and an output layer (y). The input layer takes two parameters: system error ϵ(t) and reference action r(t).
2Weight initializationThe neural network parameters kp, Ti, k1, k2, k3, and α and other adaptive parameters are initialized randomly with a normal distribution (or using the Xavier/He method). In this case, the weight matrices for the hidden layers and the bias for each layer are initialized as
W i j l N 0 , 1 n ,   b l = 0 .
3Forward passFor each l-th hidden layer, an input linear combination is computed. For the first hidden layer (error handling): z 1 = W 1 · ϵ t , r t + b 1 ,   α 1 = S m o o t h R e L U z 1
For the second hidden layer (the regulatory effect computation): z2 = W2a1 + b2, a2 = tanh(z2).
For the third hidden layer (switching parameters integration): z3 = W3a2 + b3, a3 = SmoothReLU(z3).
For the output layer (the output coordinate compute): z4 = W4a3 + b4.
4Error computationThe output error is computed as the difference between the prediction and the actual value, and a loss function, L, is applied, which could be, for example, MSE:
L = 1 n · i = 1 n y i x t i 2 .
To compute the gradients, the sensitivity function Ξ(t) is used, which takes into account the parameters’ influence on the system.
5BackpropagationBackpropagation involves computing gradients at each layer. For the output layer:
δ 4 = L y · σ z 4 ,
σ z 4 is the activation function’s derivative at the output.
For the remaining layers, starting from the third, the error is calculated based on the activation function derivative:
δ 3 = W 4 T · δ 4 · σ z 3 ,   δ 2 = W 3 T · δ 3 · σ z 2 ,   δ 1 = W 2 T · δ 2 · σ z 1 .
6Updating weights with regularizationFor each layer, the weights and biases are updated, taking into account the regularization λ. The gradients for the weights and biases are calculated as
L W l = α l 1 · δ l T + λ · W l ,   L b l = δ l .
The weights and biases are updated using the gradient descent equation:
W l = W l η l · L W l ,   b l = b l η l · L b l ,
where ηl is the dynamic training rate, depending on the change in the optimization functional:
η l = η 1 + α · Ξ t ,
where α is the training rate adaptation coefficient, and ∣Ξ(t)∣ is the change magnitude in the functional at the current step.
7Parameter’s adaptationNetwork parameters such as kp, Ti, k1, k2, k3, and α are trained via gradient descent, taking into account regularization and the sensitivity function Ξ(t). These parameters include both weights and adaptive coefficients that change based on changes in the loss function.
8Training repetitionThe training process is repeated for each batch of data or all training data, with weights and parameters updated at each step. It continues until stopping after the maximum number of epochs or the error improves.
9Performance evaluationOnce training is complete, the network is tested on test data to assess its ability to handle new data by predicting output values given nonlinearities and delays in the system.
Table 3. The training dataset fragment.
Table 3. The training dataset fragment.
NumberThe System Error ϵ(t) ValuesThe Reference Action r(t) Values
10.0110.913
460.0120.918
1150.0150.933
1730.0130.926
2090.0110.911
2560.0100.908
Table 4. The training dataset’s homogeneity, assessing results using the Fisher–Pearson and Fisher–Snedecor criteria.
Table 4. The training dataset’s homogeneity, assessing results using the Fisher–Pearson and Fisher–Snedecor criteria.
ParameterThe Fisher–Pearson Criterion MeaningThe Fisher–Snedekor Criterion, MeaningDescription
CalculatedCriticalCalculatedCritical
ϵ(t)12.99613.315.40215.98The Fisher–Pearson and Fisher–Snedecor criterium-computed values were below the thresholds, which made it possible to confirm the training dataset’s homogeneity.
r(t)13.01415.599
Table 5. The automatic parametric optimization algorithm results.
Table 5. The automatic parametric optimization algorithm results.
Number θ 1 0 θ 2 0 θ 1 * θ 2 * J0J*H(J)det(H(J))
10.10.010.2530.01638.32520.974 1.049 0.889 0.659 0.569 0.011
20.250.0250.3470.01638.26120.659 1.013 0.805 0.611 0.554 0.069
30.350.0010.4150.01259.93919.993 1.001 0.738 0.602 0.519 0.075
Table 6. Developed algorithm for conducting simulation modelling.
Table 6. Developed algorithm for conducting simulation modelling.
Modelling Step NumberModelling Step NameDescription
1Setting system parameters T o b 1 = 1.0, T o b 2 = 4.0, kob = 1.0, τob = 5.0.
Controller parameters: θ1 = 0.5, θ2 = 0.1.
Reference action: r(t) = sin(t).
2System modellingThe system block diagram creation, including the controller Gc(p, θ ¯ ) and the object Gp(p).
Adding blocks to compute ξi(t) and ξij(t).
3Sensitivity computationThe implementation of the numerical calculation of the derivatives u t θ i and 2 u t θ i θ j .
The finite difference methods or analytical solution application.
4The optimality criterion computation J = 0 a 1 · ε 2 t , θ ¯ + a 2 · u 2 t d t ,
where a1 = 1.0 and a2 = 0.5.
5Analysis of resultsThe computation of the first- and second-order sensitivity functions.
The gradient θ ¯ J and Hessian matrix H(J) computation.
Testing Sylvester’s criterion for H(J) and optimizing the θ ¯ parameters.
Table 7. Comparative analysis results for the proposed neural network training.
Table 7. Comparative analysis results for the proposed neural network training.
NumberTraining AlgorithmAchieved MSEmin ValueMinimum
Number of Epochs
Conclusions
1Developed algorithm0.0050160The most efficient learning algorithm is characterized by the minimum MSEmin value achieved in the minimum number of training epochs.
2Traditional backpropagation algorithm0.0093220The minimum MSEmin value achieved exceeds the developed algorithm’s similar indicator by 1.86 times, with an increase in the optimal number of training epochs by 1.38 times.
3Genetic algorithm0.0108240The minimum MSEmin value achieved exceeds the developed algorithm’s similar indicator by 2.16 times, with an increase in the optimal number of training epochs by 1.5 times.
4Modified inverse gradient descending method [51]0.0163250The minimum MSEmin value achieved exceeds the developed algorithm’s similar indicator by 3.26 times, with an increase in the optimal number of training epochs by 1.56 times.
5Traditional inverse gradient descending method0.0237330The minimum MSEmin value achieved exceeds the developed algorithm’s similar indicator by 4.74 times with an increase in the optimal number of training epochs by 2.07 times.
6Hybrid algorithm0.0301380The minimum MSEmin value achieved exceeds the developed algorithm’s similar indicator by 6.02 times, with an increase in the optimal number of training epochs by 2.38 times.
Table 8. Comparative analysis results.
Table 8. Comparative analysis results.
Neural Network ArchitectureMetrics for Evaluating MethodsTraining Time
AccuracyPrecisionRecallF1-ScoreAUC-ROC
Developed neural network0.9930.9911.00.9950.8251 min 12 s
Traditional RBF network [20,56]0.9920.9891.00.9940.8112 min 25 s
Jordan neural network [21,57]0.9930.9891.00.9940.7814 min 17 s
Elman neural network [22,58]0.9930.9891.00.9940.7804 min 12 s
Hopfield neural network [23,59]0.8370.8320.8350.8330.6350 min 47 s
Hamming neural network [24,60]0.8110.8060.8310.8190.6020 min 33 s
Table 9. The proposed roadmap for future research.
Table 9. The proposed roadmap for future research.
Research TitleAimActionsExpected Results
Optimization of model parameters and increased error tolerance [74].Develop methods for identifying parameters with increased accuracy and error tolerance.Study more accurate methods for estimating the controlled object parameters using machine learning and adaptive filtering methods.Reduce the model’s sensitivity to errors in parameters, increasing the modelling and control accuracy.
Develop algorithms for automatic correction of errors in parameters in real time.
Develop methods for taking into account uncertainty in model parameters.
Controller adaptation and tuning simplification [75,76].Simplify the process of tuning controller adaptation parameters and optimization coefficients for practical applications.Investigate new approaches for the automatic tuning of controller parameters (e.g., using optimization methods such as genetic algorithms or gradient descent).Simplify the tuning process and increase the method’s flexibility when applied to various control objects.
Develop universal methods for parameter optimization that can automatically adapt to system changes.
Processing complex nonlinearities [77].Develop methods for more efficient processing of complex nonlinearities in object dynamics.Study more complex neural network architectures (e.g., deep neural networks considering multilevel nonlinear dependencies).Increase the accuracy and universality of the model for systems with different nonlinear dynamics.
Implement methods considering the dynamic nature of nonlinearities, such as adaptive neural network models.
Robustness to noise and data quality [78].Improve the method’s robustness to noise and poor data quality.Develop data preprocessing methods, including improving noise filtering and eliminating sensor errors.Improve the accuracy of the method by improving the input data quality.
Use high-reliability data methods, such as data reconstruction or fragmentation methods, to improve the training datasets’ representativeness.
Reducing computational complexity [79].Optimizing computational processes and reducing the need for computational resources.Developing more efficient algorithms for adapting the training rate and regularization to speed up computations.Reduce computational costs while maintaining the high accuracy of the method.
Investigating the use of hardware accelerators (e.g., neural processors or graphics processors) to optimize computational processes.
Increasing the method generalization ability [80].Develop methods for adaptation and training transfer to apply the technique to the systems and engines’ new types.Study methods for training transfer that allow the model to be adapted to different systems and conditions.Increase the generalization ability of the method and its successful application to different types of engines and systems.
Development of universal neural network architectures suitable for different types of control objects.
The significant delays in modelling optimization [81].Develop methods for more accurate modelling of significant delays in the system.Research and implementation of methods that take into account dynamic delays and significant delays in the system.Improve model accuracy for systems with long delays.
Development of prediction methods to take into account long delays timing models’ integration to take into account phase shifts.
Simplifying the implementation of accurate systems [82].Simplify the implementing process of the method on real systems with limited computing power.Developing lighter and faster versions of the algorithm for implementation on systems with limited computing power.Increase the availability and practicality of the method for implementation in real systems with limited resources.
Investigating the possibilities of integration with existing platforms, such as microcontrollers and specialized neuro-microprocessors.
Preventing overfitting [83].Develop methods to prevent overfitting on small or heterogeneous training datasets.Use cross-validation and regularization methods to prevent overfitting.Reduce risk of overfitting and improve generalization properties of the model.
Investigate methods for generating training data and expanding the dataset to improve model robustness.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Vladov, S.; Scislo, L.; Szczepanik-Ścisło, N.; Sachenko, A.; Vysotska, V. Neural Network Method of Controllers’ Parametric Optimization with Variable Structure and Semi-Permanent Integration Based on the Computation of Second-Order Sensitivity Functions. Appl. Sci. 2025, 15, 2586. https://doi.org/10.3390/app15052586

AMA Style

Vladov S, Scislo L, Szczepanik-Ścisło N, Sachenko A, Vysotska V. Neural Network Method of Controllers’ Parametric Optimization with Variable Structure and Semi-Permanent Integration Based on the Computation of Second-Order Sensitivity Functions. Applied Sciences. 2025; 15(5):2586. https://doi.org/10.3390/app15052586

Chicago/Turabian Style

Vladov, Serhii, Lukasz Scislo, Nina Szczepanik-Ścisło, Anatoliy Sachenko, and Victoria Vysotska. 2025. "Neural Network Method of Controllers’ Parametric Optimization with Variable Structure and Semi-Permanent Integration Based on the Computation of Second-Order Sensitivity Functions" Applied Sciences 15, no. 5: 2586. https://doi.org/10.3390/app15052586

APA Style

Vladov, S., Scislo, L., Szczepanik-Ścisło, N., Sachenko, A., & Vysotska, V. (2025). Neural Network Method of Controllers’ Parametric Optimization with Variable Structure and Semi-Permanent Integration Based on the Computation of Second-Order Sensitivity Functions. Applied Sciences, 15(5), 2586. https://doi.org/10.3390/app15052586

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop