You are currently on the new version of our website. Access the old version .
Applied SciencesApplied Sciences
  • Article
  • Open Access

13 January 2026

Machine Learning and Operator-Based Nonlinear Internal Model Control Design for Soft Robotic Finger Using Robust Right Coprime Factorization

and
Department of Electrical Engineering and Computer Science, Tokyo University of Agriculture and Technology, Tokyo 184-8588, Japan
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue New Topics on System Learning and Control and Its Applications

Abstract

Currently, machine learning (ML) methods provide a practical approach to model complex systems. Unlike purely analytical models, ML methods can describe the uncertainties (e.g., hysteresis, temperature effects) that are difficult to deal with, potentially yielding higher-precision dynamics by a learning plant given a high-volume dataset. However, employing learning plants that lack explicit mathematical representations in real-time control remains challenging, namely, the model can be conversely looked at as a mapping from input data to output, and it is difficult to represent the corresponding time relationships in real applications. Hence, an ML and operator-based nonlinear control design is proposed in this paper. In this new framework, the bounded input/output spaces of the learning plant are addressed rather than mathematical dynamic formulation, which is realized by robust right coprime factorization (RRCF). While the stabilized learning plant is explored by RRCF, the desired tracking performance is also considered by an operator-based nonlinear internal model control (IMC) design. Eventually, practical application on a soft robotic finger system is conducted, which indicates the better performance of using the controlled learning plant and the feasibility of the proposed framework.

1. Introduction

In recent years, machine learning (ML) methods have attracted considerable interest from researchers, and have been widely applied in various fields like computer vision [1,2,3,4], system identification [5,6,7], robotics [8,9,10,11,12], and so on. Specifically, driven by large-scale datasets, ML offers new approaches to describe the dynamics of complex industrial systems whose analytical mathematical models are difficult to compute [13]. Given the obtained learning plant (i.e., the trained ML model), several frameworks were introduced to address the control issue. For example, a neural network (NN)-based nonlinear internal model control (IMC) structure was constructed in [14], where neural networks were selected to design the inverse model of IMC; a similar NN-based IMC structure can also be found in [15], which motivates further exploration of ML-IMC combinations; besides IMC, NN predictive control was proposed in [16], which dealt with the delay issue using recurrent NNs for nonlinear plants; an adaptive control framework was depicted in [17], where the NN played roles in both online identification and parameters adjustment; and for a multi-mode process, an LSTM-based model predictive control scheme was depicted in [18], which automatically matched different operation modes without switching strategies. However, the aforementioned studies rely on the dynamics of ML methods or objective systems, which makes the analysis of stability or tracking performance complicated. In addition, for those ML methods or systems whose dynamics are not concisely modeled, one may find it difficult to apply such control designs.
As one of the well-established nonlinear control strategies, the operator theory is implemented in this work. According to operator theory, a given nonlinear plant can be interpreted as an operator that maps from its input space towards its output space. In contrast to control schemes whose design depends on accurate system dynamics, a robust right coprime factorization (RRCF) feedback structure is established given the input/output spaces of each operator. Hence, RRCF realizes nonlinear control from the perspective of operators, namely, input–output mappings. RRCF has been practically applied in Hamiltonian systems [19], micro-hand actuator system [20], etc [21,22]. Moreover, RRCF was combined with ML methods like multi support vector regression for soft actuators in [23] and with a neural network to address the denial of service attacks in [24], where the learning plant is reliable in real applications. For learning plants that lack an explicit mathematical model and time relationships, their input–output mappings coincide with the characteristics of RRCF, which makes RRCF a potential approach to stabilize learning plants.
Motivated by the aforementioned issues, the main contributions of this work are outlined as follows:
  • An operator-based control design for machine learning models is established in the RRCF feedback structure, which ensures the robust stability of machine learning models by input–output mappings rather than explicit mathematical dynamics. Such compensation by RRCF for machine learning models provides security for practical implementation.
  • The proposed RRCF feedback system for machine learning models is inserted into a nonlinear IMC framework. Such a combination contributes to the finite tracking convergence of output, even with potential jumps or perturbation.
  • As a practical application of the proposed control design, experiments with a soft robotic finger are conducted. The comparative results indicate the feasibility and superiority of this framework.
The remainder of this paper is organized as follows. Section 2 introduces the mathematical preliminaries of RRCF. Section 3 describes the proposed control design, which includes the stabilization of the learning plant and analysis of tracking performance. Section 4 exhibits the practical experiments with the soft robotic finger. The whole work is concluded in Section 5.

2. Preliminaries

In this section, some mathematical preliminaries on nonlinear robust right coprime factorization (RRCF) will be briefly introduced. Initially, nonlinear right factorization is described as follows.
Consider the objective target plant, which is denoted by P . P is of nonlinear dynamics, and is analyzed from the perspective of an operator, i.e., P is a mapping from U to Y , which are the input space and output space, respectively. It should be noted that in the real world, such spaces are defined as associated with Banach space in the time domain. Assume two bounded subspaces in U and Y as u s and Y s , then P is said to be stable if P ( U s ) Y s . In other words, a stable operator maps bounded input space into bounded output space. However, P is potentially unstable, while indicating unbounded output space. To stabilize P , one first defines the right factorization of P as
P = N D 1
where P is factorized into two individual operators, analytically. The internal variable is defined as a quasi-state of P , denoted by ω ( t ) W s and W s is a subspace of W . Then, one can write ω ( t ) = D 1 ( u ) ( t ) , y ( t ) = N ( ω ) ( t ) . Here the requirements of N and D 1 are listed as follows:
  • N is stable from W s to Y s .
  • D 1 is potentially unstable (inherits from P ). However, D 1 is invertible between W and U , while D is stable from W s to U s .
Next, given the right factorization of P , a feedback system is constructed, which is depicted as Figure 1. To realize the stabilization of P , one needs other controllers S and R , by which the following Bezout Identity [19]
S N + R D = M 1
is satisfied. The designed S , R , and M are described as follows:
Figure 1. General scheme of right coprime factorization (RCF).
  • R maps U s to U s , i.e., e ( t ) U s , e ( t ) = R ( u ) ( t ) . R is stable and invertible. R 1 is stable.
  • S maps Y s to U s , i.e., b ( t ) U s , b ( t ) = S ( y ) ( t ) . S is stable.
  • M 1 maps W s to U s , i.e., r ( t ) = M 1 ( ω ) ( t ) . M 1 is stable and invertible, M 1 1 is stable. An operator like M 1 is defined as a unimodular operator (both M 1 and M 1 1 are stable).
Given the Bezout Identity, one can find that each of the variables e , u , ω , y , and b are stable (bounded) if the reference input r is bounded. Hence, it is called bounded input bounded output (BIBO) stable.
More generally, consider the potential uncertainty in a real-world scenario; RRCF is discussed to explore the robustness of such a feedback scheme. In Figure 2, Δ P denotes the bounded uncertainty and is factorized into Δ N as ( P + Δ P ) = ( N + Δ N ) D 1 . Then, one can rewrite the Bezout Identity as
S ( N + Δ N ) + R D = M 2
where M 2 can be found to be unimodular if S and Δ N satisfy the following inequality [19]
( S ( N + Δ N ) S N ) M 1 1 < 1
where · is the Lipschitz semi-norm, which is defined as
Q = sup T [ 0 , ) sup u , u ˜ U s u T u ˜ T | [ Q ( u ) ] T [ Q ( u ˜ ) ] T | Y B | u T u ˜ T | U B
for an nominal operator Q ( U s ) Y s , and U B , Y B denote the associated Banach spaces, respectively.
Figure 2. General scheme of robust right coprime factorization (RRCF).

3. Main Results

In this section, the basic scheme of a machine learning-based nonlinear internal control design will be described. The proposed scheme is established following the operator control theory and in the form of right coprime factorization. Moreover, the stability analysis and tracking performance of the system will be presented.

3.1. Problem Statement

Initially, the problem addressed in this paper is described here, which starts from the notations of the objective nonlinear plant, i.e.,
y 0 ( t ) = P n o m i n a l ( u ) ( t ) = P 0 ( u ) ( t )
y ^ ( t ) = P l e a r n i n g ( u ) ( t ) = ( P 0 + Δ P 0 + d 0 ) ( u ) ( t )
y ( t ) = P r e a l ( u ) ( t ) = ( P 0 + Δ P + d ) ( u ) ( t )
where P n o m i n a l denotes the mathematical dynamics (also noted as P 0 ) of the objective plant, which is generally derived from physical analysis; P l e a r n i n g serves as the model obtained by machine learning, such as support vector regression and neural network; and P r e a l is the physical plant in the real world, i.e., the target system. y 0 , y ^ , and y denote the outputs in different scenarios, respectively, and u is the input. Δ P represents the uncertainty that is difficult to analytically model or compute, which is approximated by the collected dataset and modeled as Δ P 0 using machine learning. d 0 represents the measurement perturbation that naturally occurs during data collection. d denotes the perturbation in real-world operation, as well as the mismatch between the learning plant and real-world dynamics. To describe the plant as an operator, all variables are defined in the time domain. Although d may contain a component similar to d 0 , there are assumed to be no probabilistic relations between d and d 0 .
The problems addressed in this paper can be summarized as follows:
  • In many cases, uncertainties and perturbation cannot be precisely captured, which results in a significant discrepancy between nominal models and real plants. This may affect performance (e.g., tracking accuracy). Hence, the usage of well-trained machine-learning models for control design needs to be investigated.
  • Despite their accurate approximation of real plants, learning models often do not admit a mathematical formulation and are commonly characterized as input–output relations. Therefore, ensuring closed-loop stability and satisfactory performance requires careful consideration in the control design.

3.2. Scheme of the Proposed Control Design

To address the aforementioned problems, the proposed control design can be depicted as in Figure 3, which is composed of an operator-based nonlinear internal model mechanism [25] and tracking compensation. “ML plant” and “Real plant” stand for machine learning model P l e a r n i n g and real plant P r e a l , respectively. Next, the specific details of each operator and controller will be described.
Figure 3. The proposed nonlinear internal model control design.
Given the definitions (6)–(8), the right factorization of different plants can be formulated as
y 0 ( t ) = P 0 ( u ) ( t ) = N D 1 ( u ) ( t )
y ^ ( t ) = ( P 0 + Δ P 0 + d 0 ) ( u ) ( t ) = ( N + Δ N 0 + d 0 ) D 1 ( u ) ( t )
y ( t ) = ( P 0 + Δ P + d ) ( u ) ( t ) = ( N + Δ N + d ) D 1 ( u ) ( t )
where D 1 denotes the potentially unstable component, and the uncertainty ( Δ N 0 , Δ N ) and perturbation ( d , d 0 ) that exist with N are represented. Further, the following discussion is limited to the representation of the ML plant (10). Here one can find the bounded stability of N + Δ N 0 + d 0 and N + Δ N + d given the assumption of boundedness: Δ N 0 ( · ) Δ N ¯ 0 , Δ N ( · ) Δ N ¯ , d 0 ( · ) d 0 ¯ , d ( · ) d ¯ .
In this work, the operators A and B are designed with a nominal plant. That is, by designing the proper A and B , one has to satisfy the Bezout Identity (2), which ensures the robust stability of the real plant by the unimodular operator M , though the “perturbed” Bezout Identity is likely to be unknown. Moreover, for the feedback loop of the ML plant, A 0 and B 0 offer much more freedom according to the input/output space of the ML plant, which is already known as dataset. However, such a design of A 0 and B 0 remains constrained by the Bezout Identity (2). This is aimed at the stabilization of the learning plant P l e a r n i n g , since the input/output variation is required to be within its corresponding stable space, though some perturbation or finite jumps potentially exist in the sequence. In other words, owing to the robustness and “bounded space mapping” perspective of the RRCF feedback system, the stability analysis no longer requires a precise model of the perturbation or uncertainties. In the following analysis, to validate the feasibility of the proposed design, for simplicity, A 0 and B 0 are assumed to be designed to have the same structure as A and B in this paper, i.e., A 0 N + B 0 D = A N + B D = M , leaving possibly different designs for future works. As a result, a compensated learning plant is realized by the RRCF feedback loop according to Figure 3.
Next, following the description of designed operators A , B , A 0 , and B 0 , the proposed control design will first be analyzed without considering tracking compensation, i.e., e ˜ ( t ) = r ¯ ( t ) y ˜ ( t ) . According to the principle of an internal model control mechanism, a stable controller C will be designed to satisfy
N M 1 C ( · ) = I ( · )
where I is an Identity operator and bounded. However, one will find that given the existence of uncertainty and perturbation in the ML plant, the Bezout Identity has to be rewritten as
A 0 ( N + Δ N 0 + d 0 ) + B 0 D = M ˜
where M ˜ remains unimodular if satisfying the condition (4) on the robust stability of right coprime factorization. Then, y ^ can be formulated as
y ^ ( t ) = ( N + Δ N 0 + d 0 ) M ˜ 1 C ( e ˜ ) ( t )
which results in
y ^ ( t ) e ˜ ( t ) = [ ( N + Δ N 0 + d 0 ) M ˜ 1 C I ] ( e ˜ ) ( t )
provided that ( N + Δ N 0 + d 0 ) , M ˜ 1 and C are bounded, one can write (15) as
y ^ ( t ) = e ˜ ( t ) + ϵ ( t )
where ϵ ( t ) ϵ ¯ for any bounded e ˜ ( t ) . Hence, according to the nonlinear internal model control mechanism, e ˜ ( t ) = r ¯ ( t ) y ˜ ( t ) , y ˜ ( t ) = y ( t ) y ^ ( t ) , then
r ¯ ( t ) y ( t ) = ϵ ( t )
which indicates the convergence of output of the real plant in a ball of r ¯ with its radius of ϵ ¯ . Although the replacement of the “ML plant” with the nominal plant introduces a deviation within the ball, it can practically be compensated by the tracking controller and hence has a trivial influence on the system. In contrast, in particular, the relatively smaller difference y ˜ reduces the risk of output saturation, ensuring that the system is within the robust stable region. This analysis will be validated by the experiments in the next section.
Besides the internal model control design, another controller F is selected to guarantee the tracking compensation. Specifically, F is formulated as a Proportional–Integral controller:
x e ( t ) = r ˜ ( t ) y ( t )
F ( x e ) ( t ) = K P x e ( t ) + K I 0 t x e ( τ ) d τ
then (17) can be rewritten as
r ˜ ( t ) y ( t ) = F ( x e ) ( t ) ϵ ( t ) = x e ( t )
by which the dynamics of x e ( t ) can be derived as
x ˙ e ( t ) = α x e ( t ) β ϵ ˙ ( t )
where
α = K I K P + 1 , β = 1 K P + 1
by selecting K P > 1 and K I > 0 , one can find α > 0 and β > 0 . However, with the presence of finite jumps or perturbation, ϵ ˙ in the right hand is likely to be too large and even unbounded in certain moments, which makes it difficult for stability analysis. In this work, inspired by the results of impulsive differential equations introduced in [26], the dynamics (21) are analyzed both in flow and jump intervals by defining
z ( t ) : = x e ( t ) + β ϵ ( t )
If ϵ ( t ) has a jump Δ ϵ ( t k ) : = ϵ ( t k + ) ϵ ( t k ) at t k , k = 1 , 2 , , then integrate (21) over ( t k η , t k + η ) and let η 0 :
x ˙ e d t + α x e d t + β d ϵ = Δ x e ( t k ) + α x e d t + β Δ ϵ ( t k ) = 0 ,
and the middle integral vanishes as the interval shrinks, which results in
Δ x e ( t k ) : = x e ( t k + ) x e ( t k ) = β Δ ϵ ( t k )
Introducing (25) into the derivation of (23), one can find Δ z ( t k ) = Δ x e ( t k ) + β Δ ϵ ( t k ) = 0 , which indicates that z is continuous at jumps. During flow intervals (between jumps),
z ˙ ( t ) = x ˙ e ( t ) + β ϵ ˙ ( t ) = α x e ( t ) = α ( z ( t ) β ϵ ( t ) )
for any t t 0 , this leads to
z ( t ) = e α ( t t 0 ) z ( t 0 ) + α β t 0 t e α ( t s ) ϵ ( s ) d s
consequently, the convergence of z ( t ) is obtained as
| z ( t ) | e α ( t t 0 ) | z ( t 0 ) | + β 1 e α ( t t 0 ) ϵ ¯
and since x = z β ε , one can write the limits as
lim sup t | z ( t ) | β ϵ ¯ lim sup t | x e ( t ) | 2 β ϵ ¯
Hence, by tuning K P , K I , the tracking error x e ( t ) will be uniformly ultimately bounded within trivial ϵ ¯ ; moreover, in particular, if ϵ ( t ) settles as a constant, i.e., ϵ ( t ) ϵ , then
z ( t ) β ϵ and x e ( t ) = z ( t ) β ϵ ( t ) 0 .
by which (possibly discontinuous) convergence of x e ( t ) to the origin can be obtained. Moreover, the negative impact of jumps and perturbation can be addressed intrinsically by the learning algorithm as well. Specifically, for example, in support vector regression (SVR), one may select an appropriate kernel and tune the ε -insensitive loss to attenuate perturbation, whereas in neural networks, choosing smooth activation with bounded Lipschitz constants ensures continuity and well-behaved derivatives of the output.
Theoretically, the superiority of this framework can be emphasized as follows. First, the learning plant is stabilized by RRCF, from the perspective of input–output mappings. Without explicit mathematical dynamics and corresponding time relationships of ML plants, the RRCF feedback loop of the represented learning plant (10) is found to be BIBO stable even in the presence of uncertainty and perturbation. In addition, compared to nominal plants, the learning plants are much closer to the dynamics in the real world, which contributes to better tracking performance by nonlinear IMC, especially with presence of output saturation. Moreover, another improvement compared to [27] is related to the individual RRCF feedback loop for P l e a r n i n g and P r e a l , which stabilizes the plant by different input u 0 and u , respectively. However, the framework in [27] uses the same input for P l e a r n i n g and P r e a l , which potentially results in the instability of P r e a l due to the mismatch and real-time perturbation between them. Hence, the framework depicted in Figure 3 is recommended. In what follows, the theoretical analysis of the proposed control design will be supported by an experimental evaluation.

4. Experimental Application on Soft Robotic Finger

In this section, the proposed machine learning and operator-based nonlinear internal model control design is applied to a pneumatic soft robotic finger. The details of structure, experimental process, controller design, and comparative results are exhibited.

4.1. Physical Structure and Mathematical Model of Soft Robotic Finger

Initially, the physical structure and mathematical model (nominal plant) are introduced. The soft robotic finger is composed of semi-cylindrical rubber, which was designed in [28] to realize bending. Specifically, according to the structure model shown in Figure 4, several similar bellow units can be found in the round side, each of which is divided into channel and chamber. The difference in height provides enough space to bend, which makes the round side more extensible than the flat side. The bending soft robotic finger with different input pressures in real-world application can be found in Figure 5.
Figure 4. Physical structure of the soft robotic finger: (a) structural model; (b) cross section.
Figure 5. Bending soft robotic finger with different input pressures.
To derive the mathematical model of the soft robotic finger, the variables and parameters in Figure 4 are introduced. p (kPa) denotes the input pressure; θ (rad) is the output bending angle. L 0 serves as the initial length of the finger and Δ L is the change during bending. l denotes the length of each unit, while r 0 is the inner radius of each unit and κ is the thickness of the rubber. According to the results in [28], the equations of moments are derived from the integral of elastic force and pressure within radius [ 0 , r ] and angle [ 0 , φ ] , by which the mathematical model of the soft robotic finger can be written as
θ = L 0 κ 1 + 2 R + κ π γ p 2 R κ π
γ = 4 r 0 3 + 3 π r 0 2 κ 6 r 0 κ E 0
where E 0 is the Young’s modulus and R = r 0 + κ 2 . It should be noted that the input pressure and output angle are physically constrained within p 0 , 60 , θ 0 , π . In the following analysis, Equation (31) will serve as the nominal plant for control design in comparative experiments.

4.2. Experimental Set up and Process

The experimental set up for the soft robotic finger and details of the process are introduced here. A visible overview of the experimental set up is exhibited in Figure 6, while the detailed process is shown in Figure 7.
Figure 6. Overview of experimental set up.
Figure 7. Experimental process.
Accordingly, one can find that the soft robotic finger and camera are placed on a vibration-isolated platform, which protects them from ground-induced vibration and perturbation. The air pressure is provided by a compressor, which is connected to two regulators. Specifically, regulator 1 is selected to filter out oil and dust, while the other (regulator 2) is involved in security by monitoring the pressure level. To realize the automatic control on the soft robotic finger, an electro-pneumatic regulator is selected, which is driven by an external power supply. The output pressure of the electro-pneumatic regulator is proportional to the input current, which is controlled by PC through a data acquisition (DAQ) device and a voltage–current converter. Finally, images of the soft robotic finger are collected by a camera, which contributes to the calculation of bending angle.

4.3. Neural Network Training and Controllers Design

To realize the proposed machine learning and operator-based nonlinear internal model control, a single-input single-output three-layer feedforward neural network trained via the back propagation (BP) algorithm is employed in this work for the establishment of the learning plant. According to the nominal model (31), the input–output relationship of the considered soft robotic finger is static, which makes it suitable to be approximated by a neural network. Firstly, the dataset is collected by setting the input pressure u ( t ) = 30 + 30 sin ( π 20 t ) , t [0 s, 100 s] and the sampling time d t = 0.1 s. As a result, 1000 samples are collected; 60% of them are used for training and 40% for validation.
Moreover, the structure of the neural network is depicted in Figure 8, where x and y denote the input and output of the plant, respectively. In order to achieve a good balance between approximation accuracy and computational complexity, based on multiple training trials, the rest of the neural network is configured as follows: k = 50 neurons are selected in hidden layer, with weights v k , w k ; γ k and ρ denote the threshold of each neuron; the learning rate is set as 0.3, with 60 training iterations and a stopping threshold of 1 × 10 5 ; the activation function is selected as
f ( x ) = 1 1 + e x
and the loss functions are
J t r a i n = 1 N t r a i n j = 1 N t r a i n ( y ^ j y j ) 2
J v a l = 1 N v a l j = 1 N v a l ( y ^ j y j ) 2
where N t r a i n and N v a l denote the total number of training and validation samples; in addition, a coefficient of determination R 2 is formulated as follows to evaluate the precision of neural network prediction:
R 2 = 1 j = 1 N v a l ( y ^ j y j ) 2 j = 1 N v a l ( y j y ¯ ) 2
where y ¯ = 1 N v a l j = 1 N v a l y j .
Figure 8. Structure of neural network for soft robotic finger.
Finally, convergence of the loss functions during neural network training is shown in Figure 9, and the comparative results between the nominal plant, learning plant, and real plant are exhibited in Figure 10.
Figure 9. Convergence of the loss functions.
Figure 10. Outputs of nominal plant (dataset), learning plant (network output), and real plant.
It can clearly be observed from Figure 10 that the soft robotic finger performs hysteresis in the real world, which makes the nominal plant with output saturation less accurate to describe the dynamics. In contrast, one can see that the learning plant is closer to the real plant, with a precision of R 2 = 0.9178 . This indicates the applicability of the obtained learning plant for nonlinear internal model control and is expected to result in better tracking performance.
For the next step, following the scheme described in Section 3, the controllers A , A 0 , B , B 0 and C will be designed. According to the operator theory in the RCF feedback system, the mathematical model (31) can be factorized into
N ( ω ) ( t ) = L 0 κ 1 + 2 R + κ π ω ( t ) 2 R + κ π
D 1 ( u ) ( t ) = γ · u ( t )
To satisfy the Bezout Identity, A and B can be designed as
A ( y 0 ) ( t ) = 1 K L 0 ( 2 R + κ π ) κ y 0 ( t ) + L 0 + 2 R + κ π
B ( u ) ( t ) = γ K · u ( t )
where K is an designed parameter and is selected as K = 1.25 in this work. As a result, one can find A N + B D = I based on (37)–(39), which suggests the robust stability of the RRCF feedback system. Moreover, to coincide with the concept of internal model control, C is designed as C = N 1 .
Additionally, to explore the feasibility of the neuron network based on the RRCF feedback loop, i.e., the flow from e ˜ ( t ) towards y ^ in Figure 3, a simulation verification is conducted. The simulated results are exhibited in Figure 11, where stable outputs are obtained from bounded external input e ˜ ( t ) = 1.4 + 0.8 / e 9 t + 90 . Hence, the neuron network-based RRCF feedback loop is considered safe for further practical implementation.
Figure 11. Simulated results of neuron network-based output within RRCF.

4.4. Experimental Results

Comparative experiments are conducted between the machine learning-based IMC system (the proposed method) and the IMC system in [25] with PI/PD controllers. The reference input is set as
r ˜ ( t ) = 1.4 + 0.8 e 9 t + 90
and the total experiment period is 20 s. Moreover, according to the aforementioned parameter ranges of (22) and multiple experimental trials conducted under various scenarios, specific parameters selected for different controllers are listed in Table 1.
Table 1. Parameters selection in different scenarios.
The visible results from different scenarios can be found in Figure 12, Figure 13, Figure 14 and Figure 15. It is observed in Figure 12 that, when the target value is relatively lower, each method produces outputs that track the target. In contrast, despite implementing the PI/PD controller, the outputs of a normal IMC-based system cannot converge to the higher reference, due to the output saturation and larger difference between the internal model and the real plant. This claim is further supported by Figure 13. The tracking errors in different scenarios converge to zero when the reference is r ˜ ( · ) = 1.4 rad, while the ones using normal IMC perform a tracking error of 0.2 for reference r ˜ ( · ) = 2.2 rad. Such a difference in tracking errors verifies the essential role of the machine learning plant in this framework.
Figure 12. Outputs comparison in different scenarios.
Figure 13. Tracking errors in different scenarios.
Figure 14. Outputs of internal model in different scenarios.
Figure 15. Input pressure in different scenarios.
The outputs of the internal model y ^ can be found in Figure 14. The saturation after 10 s also indicates the negative influence on the control performance when using an imprecise mathematical model. In other words, this suggests the necessity of implementing the machine learning plant to prevent avoidable saturation, as well as the effectiveness of inserting a machine learning plant into an IMC framework. Finally, the input pressures in different scenarios are exhibited in Figure 15. Besides the neural network, a support vector regression (SVR)-based machine learning plant is implemented into the proposed framework, and the experimental results are exhibited in Figure 16, which further supports the feasibility of the ML-based IMC design.
Figure 16. Outputs comparison in different scenarios while using SVR.

5. Conclusions

In this work, the practical implementation of machine learning models in a real-world control system is addressed. Although the obtained learning plant lacks mathematical dynamic formulation, the bounded input/output spaces are ensured using an RRCF feedback structure. In addition to stability, the tracking performance is investigated under a nonlinear IMC architecture. The proposed framework is then applied to a soft robotic finger system. According to the results, one can find superior performance when using ML-based IMC, since a more precise learning model avoids negative saturation. Such a finding also implies the feasibility of the proposed framework and the requirement of learning plants in some cases.
Moreover, the proposed framework is considered to be of far-reaching significance. By leveraging RRCF, the learning plants can be inserted as long as they satisfy the conditions on input/output spaces, even with the existence of perturbation or finite jumps. Future research will address ML-based enhancement on controller design, considering stability and security in practice.

Author Contributions

Conceptualization, M.D.; Methodology, Z.A.; Software, Z.A.; Validation, Z.A.; Formal analysis, Z.A.; Investigation, Z.A.; Resources, Z.A.; Data curation, Z.A.; Writing—original draft, Z.A.; Writing—review & editing, M.D.; Visualization, Z.A.; Supervision, M.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Meng, L.; Hirayama, T.; Oyanagi, S. Underwater-drone with panoramic camera for automatic fish recognition based on deep learning. IEEE Access 2018, 6, 17880–17886. [Google Scholar] [CrossRef]
  2. Yue, X.; Meng, L. YOLO-MSA: A multi-scale stereoscopic attention network for empty-dish recycling robots. IEEE Trans. Instrum. Meas. 2023, 72, 2528014. [Google Scholar] [CrossRef]
  3. Ge, Y.; Li, Z.; Yue, X.; Li, H.; Li, Q.; Meng, L. IoT-based automatic deep learning model generation and the application on empty-dish recycling robots. Internet Things 2024, 25, 101047. [Google Scholar] [CrossRef]
  4. Mahadevkar, S.; Khemani, B.; Patil, S.; Kotecha, K.; Vora, D.; Abraham, A.; Gabralla, L. A review on machine learning styles in computer vision—techniques and future directions. IEEE Access 2022, 10, 107293–107329. [Google Scholar] [CrossRef]
  5. Chen, S.; Billings, S.; Grant, P. Non-linear system identification using neural networks. Int. J. Control 1990, 51, 1191–1214. [Google Scholar] [CrossRef]
  6. Wang, J.; Chen, Y. A fully automated recurrent neural network for unknown dynamic system identification and control. IEEE Trans. Circuits Syst. I Regul. Pap. 2006, 53, 1363–1372. [Google Scholar] [CrossRef]
  7. Chiuso, A.; Pillonetto, G. System identification: A machine learning perspective. Annu. Rev. Control. Robot. Auton. Syst. 2019, 2, 281–304. [Google Scholar] [CrossRef]
  8. Patino, H.; Carelli, R.; Kuchen, B. Neural networks for advanced control of robot manipulators. IEEE Trans. Neural Netw. 2002, 13, 343–354. [Google Scholar] [CrossRef] [PubMed]
  9. Lee, T.; Christopher, J. Adaptive Neural Network Control of Robotic Manipulators; World Scientific: Singapore, 1998. [Google Scholar]
  10. Dikici, O.; Ghignone, E.; Hu, C.; Baumann, N.; Xie, L.; Carron, A. Learning-based on-track system identification for scaled autonomous racing in under a minute. IEEE Robot. Autom. Lett. 2025, 10, 1984–1991. [Google Scholar] [CrossRef]
  11. Soori, M.; Arezoo, B.; Dastres, R. Artificial intelligence, machine learning and deep learning in advanced robotics, a review. Cogn. Robot. 2023, 3, 54–70. [Google Scholar] [CrossRef]
  12. Semeraro, F.; Griffiths, A.; Cangelosi, A. Human–robot collaboration and machine learning: A systematic review of recent research. Robot. Comput.-Integr. Manuf. 2023, 79, 102432. [Google Scholar] [CrossRef]
  13. Wang, D.; Dang, G. Fuzzy recurrent stochastic configuration networks for industrial data analytics. IEEE Trans. Fuzzy Syst. 2024, 33, 1178–1191. [Google Scholar] [CrossRef]
  14. Rivals, I.; Personnaz, L. Nonlinear internal model control using neural networks: Application to processes with delay and design issues. IEEE Trans. Neural Netw. 2000, 11, 80–90. [Google Scholar] [CrossRef] [PubMed]
  15. Bonassi, F.; Scattolini, R. Recurrent neural network-based internal model control design for stable nonlinear systems. Eur. J. Control 2022, 65, 100632. [Google Scholar] [CrossRef]
  16. Huang, J.; Lewis, F. Neural-network predictive control for nonlinear dynamic systems with time-delay. IEEE Trans. Neural Netw. 2003, 14, 377–389. [Google Scholar] [CrossRef]
  17. Zhou, X.; Wang, W.; Shi, Y. Neural network/PID adaptive compound control based on RBFNN identification modeling for an aerial inertially stabilized platform. IEEE Trans. Ind. Electron. 2024, 71, 16514–16522. [Google Scholar] [CrossRef]
  18. Huang, K.; Wei, K.; Li, F.; Yang, C.; Gui, W. LSTM-MPC: A deep learning based predictive control method for multimode process control. IEEE Trans. Ind. Electron. 2023, 70, 11544–11554. [Google Scholar] [CrossRef]
  19. Bu, N.; Zhang, Y.; Li, X.; Chen, W.; Jiang, C. Passive robust control for uncertain Hamiltonian systems by using operator theory. CAAI Trans. Intell. Technol. 2022, 7, 594–605. [Google Scholar] [CrossRef]
  20. Bu, N.; Zhang, Y.; Li, X. Robust tracking control for uncertain micro hand actuator with Prandtl-Ishlinskii hysteresis. Int. J. Robust Nonlinear Control 2023, 33, 9391–9405. [Google Scholar] [CrossRef]
  21. Katsurayama, Y.; Deng, M.; Jiang, C. Operator-based experimental studies on nonlinear vibration control for an aircraft vertical tail with considering low-order modes. Trans. Inst. Meas. Control 2015, 38, 1421–1433. [Google Scholar] [CrossRef]
  22. Takahashi, K.; Deng, M. Nonlinear sensorless cooling control for a Peltier actuated aluminum plate thermal system. In Proceedings of the 2013 International Conference on Advanced Mechatronic Systems, Luoyang, China, 25–27 September 2013; pp. 1–6. [Google Scholar]
  23. Usami, T.; Deng, M. Applying an MSVR method to forecast a three-degree-of-freedom soft actuator for a nonlinear position control system: Simulation and experiments. IEEE Syst. Man Cybern. Mag. 2022, 8, 61–69. [Google Scholar] [CrossRef]
  24. An, Z.; Deng, M. Learning and operator based control system design for soft robotic finger with denial of service attack and input hysteresis using right coprime factorization. In Proceedings of the 6th International Conference on Industrial Artificial Intelligence, Shenyang, China, 21–24 August 2024; pp. 1–6. [Google Scholar]
  25. Jin, G.; Deng, M. Non-linear forced vibration control for a vertical wing-plate with sustainable perturbation by using internal model control mechanism. Int. J. Control 2023, 96, 1543–1550. [Google Scholar] [CrossRef]
  26. Lakshmikantham, V.; Bainov, D.; Simeonov, P. Theory of Impulsive Differential Equations; World Scientific: Singapore, 1989. [Google Scholar]
  27. An, Z.; Deng, M.; Morohoshi, Y. Right coprime factorization-based simultaneous control of input hysteresis and output disturbance and its application to soft robotic finger. Electronics 2024, 13, 2025. [Google Scholar] [CrossRef]
  28. Fujita, K.; Deng, M.; Wakimoto, S. A miniature pneumatic bending rubber actuator controlled by using the PSO-SVR-based motion estimation method with the generalized Gaussian kernel. Actuators 2017, 6, 6. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Article metric data becomes available approximately 24 hours after publication online.