You are currently viewing a new version of our website. To view the old version click .
Actuators
  • Article
  • Open Access

6 July 2023

LSTM-CNN Network-Based State-Dependent ARX Modeling and Predictive Control with Application to Water Tank System

,
and
1
School of Automation, Central South University, Changsha 410083, China
2
Engineering Training Center, Hunan Institute of Engineering, Xiangtan 411101, China
3
College of Mechanical and Vehicle Engineering, Hunan University, Changsha 410082, China
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Learning Control, Fault Diagnosis, and Actuator Applications of Complex Networked Systems

Abstract

Industrial process control systems commonly exhibit features of time-varying behavior, strong coupling, and strong nonlinearity. Obtaining accurate mathematical models of these nonlinear systems and achieving satisfactory control performance is still a challenging task. In this paper, data-driven modeling techniques and deep learning methods are used to accurately capture a category of a smooth nonlinear system’s spatiotemporal features. The operating point of these systems may change over time, and their nonlinear characteristics can be locally linearized. We use a fusion of the long short-term memory (LSTM) network and convolutional neural network (CNN) to fit the coefficients of the state-dependent AutoRegressive with the eXogenous variable (ARX) model to establish the LSTM-CNN-ARX model. Compared to other models, the hybrid LSTM-CNN-ARX model is more effective in capturing the nonlinear system’s spatiotemporal characteristics due to its incorporation of the strengths of LSTM for learning temporal characteristics and CNN for capturing spatial characteristics. The model-based predictive control (MPC) strategy, namely LSTM-CNN-ARX-MPC, is developed by utilizing the model’s local linear and global nonlinear features. The control comparison experiments conducted on a water tank system show the effectiveness of the developed models and MPC methods.

1. Introduction

MPC is an effective control method developed in industrial practice. It can adequately deal with the problems of multivariable, multi-constraint, strong coupling, and strong nonlinearity existing in actual industrial process objects; it has been a widespread in academia and industry and resulted in numerous theoretical research and application results [1,2]. The MPC method predicts the system’s future behavior based on the controlled object’s dynamic model to achieve the goal of optimal control. Therefore, the predictive model’s capacity to describe the system has a substantial effect on the MPC controller’s control performance. Due to the fact that some key parameters of complex nonlinear systems cannot be determined or obtained, it is difficult to obtain a physical model that accurately represents the system’s nonlinear dynamic characteristics by establishing differential or difference equations via Newtonian mechanics analysis or the Lagrange energy method [3]. Therefore, data-driven identification is a highly effective modeling approach for realizing the MPC of nonlinear systems. This method uses input and output data to construct the system model and does not need to analyze the complex interrelationship between the various physical variables of the system [4,5].
When designing predictive control algorithms based on the identified nonlinear system models, using segmented linearization [6] or local linearization [7] models can simplify the controller’s design but not adequately represent the complex system’s nonlinear dynamic characteristics, which will affect the controller’s control performance. To address the accuracy loss issue caused by linearization models, direct use of nonlinear models such as bilinear models [8,9], Volterra series models [10,11], and neural network models [12,13] can provide a better description of the system’s nonlinear dynamic characteristics. However, predictive control algorithms based on these nonlinear models need to solve non-convex optimization problems online, which may increase the computational burden and not guarantee feasible solutions. To address these issues, in recent years, many scholars have started studying combined models, including the Hammerstein model [14], Wiener model [15], Hammerstein-Wiener model [16,17], and SD-ARX model [18,19,20,21,22,23,24,25]. Among them, the SD-ARX model outperforms the Hammerstein and Wiener models in capturing the nonlinear system’s dynamic properties. The SD-ARX model uses state-dependent coefficients to represent the system’s nonlinear dynamic properties, and its local linear and global nonlinear structural characteristics make it easy to design MPC controllers. In recent years, SD-ARX models obtained by fitting model coefficients using RBF neural networks [18,19,20,21,22,23], deep belief networks [24], and wavelet neural networks [25] have been widely employed for complex industrial system modeling and control. However, these neural networks belong to feedforward neural networks, and their data information can only be transmitted forward. The connection between network nodes at each layer does not form a cycle, so their ability to describe nonlinear systems is limited.
With the fast advancement of AI technology, deep learning has attained success in numerous fields [26]. Its core idea is to learn the high-dimensional features of data via multi-level nonlinear transformation and a back-propagation algorithm. LSTM and CNN are popular deep learning neural networks, but their design objectives and application scenarios are different. LSTM is primarily employed for time series modeling, has strong memory ability and long-term dependence, and has achieved great success in natural language processing [27,28,29], speech recognition [30,31,32], time series prediction [33,34,35,36], etc. Because the LSTM neural network introduces a gating mechanism, it successfully solves the issue of gradient vanishing and exploding in a standard recurrent neural network (RNN), allowing it to process long sequence data more efficiently, which is conducive to nonlinear systems modeling. For example, Wu et al. [37] developed a dropout LSTM and co-teaching LSTM strategy for nonlinear systems. Terzi et al. [38] developed an MPC method based on the LSTM network and carried out numerical tests on the pH reactor simulator. Zarzycki et al. [39] developed a model prediction algorithm based on the LSTM network that has achieved good modeling accuracy and control effects by using online advanced tracking linearization. Although these modeling methods based on the LSTM neural network can properly learn the time dimension features in nonlinear system data, their ability to learn the spatial dimension features is limited, which affects the accuracy of system modeling. In addition, the validation of these methods has been proved by numerical simulation only and has not been applied to real objects.
The CNN can autonomously learn the spatial features of input data via convolution and pooling operations and is mainly used in computer vision [40,41,42,43]. Therefore, the composite neural network consisting of LSTM and CNN can effectively learn the spatiotemporal features of data and improve the accuracy of complex nonlinear modeling. Its research mainly focuses on emotion analysis [44,45,46], text classification [47,48,49], and electricity forecasting [50,51], but has not been found to be used for modeling and MPC of real industrial objects.
This article established the LSTM-ARX and LSTM-CNN-ARX models to represent a category of smooth nonlinear systems whose operating point may change over time and whose nonlinear characteristics can be locally linearized by using LSTM or/and CNN to fit the SD-ARX model’s coefficients. Furthermore, according to the pseudo-linear structure of the SD-ARX model, which exhibits both local linearization and global nonlinearity, two model predictive controllers have been designed, namely LSTM-ARX-MPC and LSTM-CNN-ARX-MPC. To evaluate the performance of these models and control algorithms, a real-time control comparative study was conducted on the commonly used water tank system in industrial process control. The outcomes showed that the developed models and MPC algorithms are effective and feasible. Particularly, the LSTM-CNN-ARX-MPC demonstrates excellent comprehensive control performance by leveraging the strengths of LSTM for learning temporal dimension characteristics and CNN for learning spatial dimension characteristics, enabling it to efficiently and accurately learn the nonlinear system’s spatiotemporal characteristics from large volumes of data. This article’s major contributions are summarized below.
(1)
The LSTM-ARX and LSTM-CNN-ARX models are proposed to describe the system’s nonlinear features.
(2)
The predictive controller was developed using the model’s pseudo-linear structure features.
(3)
Control comparison experiments were conducted on the water tank system, which is a commonly used industrial process control device, to validate the efficiency of the developed models and control algorithms. To our knowledge, there are currently no reports on using deep learning algorithms for nonlinear system modeling and the real-time control of actual industrial equipment. This study demonstrates how to establish deep learning-related models for nonlinear systems, design the MPC algorithms, and achieve real-time control rather than only doing a numerical simulation as in the relevant literature.
The article’s structure is as follows: Section 2 describes the related work. Section 3 studies three combination models. Section 4 designs the model-based predictive controllers. Section 5 presents the results of real-time control comparative experiments on the water tank system. Section 6 summarizes the research content.

3. Hybrid Models

This section describes three combination models, namely the LSTM-ARX model, CNN-ARX model, and LSTM-CNN-ARX model.

3.1. LSTM-ARX Model

LSTM neural networks utilize recurrent connections that allow information to flow in a directed cycle, enabling them to effectively capture the temporal dependencies in data during modeling. Unlike feedforward neural networks, which only propagate information forward, LSTM can retain and utilize information from previous time steps, enabling them to learn and represent long-term dependencies in the input sequences. This property makes LSTM particularly well-suited for building sequential models of nonlinear systems. Simultaneously, a gating mechanism is introduced in LSTM networks to control the flow and discarding of historical information and input characteristics, which solves the long-term dependence problem in simple RNN networks. In the actual modeling process, single-layer LSTM is usually unable to express complex temporal nonlinear characteristics, so a series approach is often adopted to stack multiple LSTM layers to deepen the network and bolster modeling capabilities. The LSTM-ARX model can be obtained by approximating the function coefficient of the model (3) with LSTM. It combines the benefits of LSTM for handling long sequence data with the SD-ARX model’s nonlinear expression capabilities to fully represent the nonlinear system’s dynamic features. Figure 3 describes the LSTM-ARX model’s architecture and its expression below:
{ y ( a ) = ρ 0 ( ε ( a 1 ) ) + i = 1 s y ρ y , i ( ε ( a 1 ) ) y ( a i ) + j = s d s u + s d 1 ρ u , j ( ε ( a 1 ) ) u ( a j ) + θ ( a ) [ ρ 0 ( ε ( a 1 ) )   , ρ y , i ( ε ( a 1 ) ) , ρ u , j ( ε ( a 1 ) ) ] = W ( f ( h d n ( a ) W h y + b y ) ) + b h l r ( a ) = σ 4 ( h l r 1 ( a ) W o x r + h l 1 r ( a ) W o h r + b o r ) σ 5 ( c l r ( a ) ) , r = 2 , 3 , , n c l r ( a ) = σ 4 ( h l r 1 ( a ) W f x r + h l 1 r ( a ) W f h r + b f r ) c l 1 r ( a ) + σ 4 ( h l r 1 ( a ) W i x r + h l 1 r ( a ) W i h r + b i r ) σ 5 ( h l r 1 ( a ) W c x r + h l 1 r ( a ) W c h r + b c r ) h l 1 ( a ) = σ 4 ( ε l W o x 1 + h l 1 1 ( a ) W o h 1 + b o 1 ) σ 5 ( c l 1 ( a ) ) ,   l = 1 , 2 , , d   c l 1 ( a ) = σ 4 ( ε l W f x 1 + h l 1 1 ( a ) W f h 1 + b f 1 ) c l 1 1 ( a ) + σ 4 ( ε l W i x 1 + h l 1 1 ( a ) W i h 1 + b i 1 ) σ 5 ( ε l W c x 1 + h l 1 1 ( a ) W c h 1 + b c 1 ) h 0 k ( a ) = 0 ,   c 0 k ( a ) = 0 , k = 1 , 2 , , n ; ε ( a 1 ) = [ ε 1 , , ε d ] T
where y ( a ) represents output; u ( a ) represents input; θ ( a ) represents white noise; s y s u , d are model’s orders; s d denotes time delay; ε ( a 1 ) represents the input data at time a , which may correspond to the system’s input or/and output; ρ 0 ( ε ( a 1 ) ) , ρ y , i ( ε ( a 1 ) ) , and ρ u , j ( ε ( a 1 ) ) represent the model’s function coefficient; n denotes the number of hidden layers; h l r ( a ) and c l r ( a ) represent the hidden layer state and cell state of the r -th hidden layer of time step l at time a ; σ 4 ( ) , σ 5 ( ) , and f ( ) represent activation functions (e.g., Hard Sigmoid and Tanh) that are employed to improve the model’s capability in capturing the system’s nonlinearity; { b f k , b i k , b o k , b c k | k = 1 , , n } , b y , and b represent offsets; { W f x k , W f h k , W i x k , W i h k , W o x k , W o h k , W c x k , W c h k | k = 1 , , n } , W h y , and W represent weight coefficients.
Figure 3. The LSTM-ARX model’s architecture.

3.2. CNN-ARX Model

CNN has strong nonlinear feature extraction capabilities. The CNN-ARX model is derived by fitting the SD-ARX model’s coefficients to CNN. This model combines the ability of CNN’s local feature extraction and the SD-ARX model’s expression. Compared to traditional neural networks, CNN has the characteristics of local connections, weight sharing, pooling operations, etc., which can further decrease the complexity and overfitting risks of the model and enable the model to possess a certain level of robustness and fault tolerance. Figure 4 describes the CNN-ARX model’s architecture and its expression below:
{ y ( a ) = ρ 0 ( ε ( a 1 ) ) + i = 1 s y ρ y , i ( ε ( a 1 ) ) y ( a i ) + j = s d s u + s d 1 ρ u , j ( ε ( a 1 ) ) u ( a j ) + θ ( a ) [ ρ 0 ( ε ( a 1 ) ) , ρ y , i ( ε ( a 1 ) ) , ρ u , j ( ε ( a 1 ) ) ] = W 1 ( f 1 ( x ˜ n 1 ( a ) W x y + b x ) ) + b 1 x ˜ n 1 ( a ) = σ 3 ( x ^ j 1 n 1 ( a ) ) ;   x ^ j 1 n 1 ( a ) = σ 2 ( x j 1 n 1 ( a ) ) x j 1 l 1 ( a ) = σ 1 ( i 1 = 1 g 1 x i 1 l 1 1 ( a ) W i 1 j 1 l 1 + b j 1 l 1 ) , 1 l 1 n 1 x 0 ( a ) = ε ( a 1 ) = [ ε 1 , , ε d ] T
where x 0 ( a ) represents input; σ 1 ( ) and f 1 ( ) represent activation functions; σ 2 ( ) and σ 3 ( ) indicate pooling operation and flattening operation; x ˜ n 1 ( a ) denotes the one-dimensional vector obtained by flattening the output feature maps x ^ j 1 n 1 ( a ) ; { W i 1 j 1 l 1 | l 1 = 1 , , n 1 } , W x y , and W 1 represent weight coefficients; { b j 1 l 1 | l 1 = 1 , , n 1 } , b x , and b 1 represent the offsets.
Figure 4. The CNN-ARX model’s architecture.

3.3. LSTM-CNN-ARX Model

The LSTM-CNN-ARX model is obtained by combining LSTM and CNN to approximate SD-ARX model coefficients, as shown in Figure 5. First, employ LSTM for learning the temporal characteristics of the input data, followed by the utilization of CNN to capture the spatial characteristics, and finally calculate the SD-ARX model’s correlation coefficients through a full connection layer to obtain the LSTM-CNN-ARX model. This model integrates the spatiotemporal characteristic extraction ability of LSTM-CNN with the expressive power of the SD-ARX model; this approach effectively captures the nonlinear system’s dynamic features. The LSTM-CNN-ARX model’s mathematical expression is obtained from Formulas (6) and (7), as follows:
{ y ( a ) = ρ 0 ( ε ( a 1 ) ) + i = 1 s y ρ y , i ( ε ( a 1 ) ) y ( a i ) + j = s d s u + s d 1 ρ u , j ( ε ( a 1 ) ) u ( a j ) + θ ( a ) [ ρ 0 ( ε ( a 1 ) ) , ρ y , i ( ε ( a 1 ) ) , ρ u , j ( ε ( a 1 ) ) ] = W 1 ( f 1 ( x ˜ n 1 ( a ) W x y + b x ) ) + b 1 x ˜ n 1 ( a ) = σ 3 ( x ^ j 1 n 1 ( a ) ) ;   x ^ j 1 n 1 ( a ) = σ 2 ( x j 1 n 1 ( a ) ) x j 1 l 1 ( a ) = σ 1 ( i 1 = 1 g 1 x i 1 l 1 1 ( a ) W i 1 j 1 l 1 + b j 1 l 1 ) , 1 l 1 n 1 x 0 ( a ) = h l n ( a ) h l r ( a ) = σ 4 ( h l r 1 ( a ) W o x r + h l 1 r ( a ) W o h r + b o r ) σ 5 ( c l r ( a ) ) , r = 2 , 3 , , n c l r ( a ) = σ 4 ( h l r 1 ( a ) W f x r + h l 1 r ( a ) W f h r + b f r ) c l 1 r ( a ) + σ 4 ( h l r 1 ( a ) W i x r + h l 1 r ( a ) W i h r + b i r ) σ 5 ( h l r 1 ( a ) W c x r + h l 1 r ( a ) W c h r + b c r ) h l 1 ( a ) = σ 4 ( ε l W o x 1 + h l 1 1 ( a ) W o h 1 + b o 1 ) σ 5 ( c l 1 ( a ) ) ,   l = 1 , 2 , , d   c l 1 ( a ) = σ 4 ( ε l W f x 1 + h l 1 1 ( a ) W f h 1 + b f 1 ) c l 1 1 ( a ) + σ 4 ( ε l W i x 1 + h l 1 1 ( a ) W i h 1 + b i 1 ) σ 5 ( ε l W c x 1 + h l 1 1 ( a ) W c h 1 + b c 1 ) h 0 k ( a ) = 0 ,   c 0 k ( a ) = 0 , k = 1 , 2 , , n ; ε ( a 1 ) = [ ε 1 , , ε d ] T
Figure 5. The LSTM-CNN-ARX model’s architecture.
The aforementioned deep learning-based SD-ARX models adopt data-driven techniques for modeling. First, select appropriate and effective input and output data as identification data for modeling. Second, based on the complexity of the identified system, select the model’s initial orders, the network’s layer count, and the node count per layer. Third, choose activation functions (such as Sigmoid, Tanh, and Relu) and optimizers (such as Adam and SGD) and configure other hyperparameters of the model. Fourth, train the model and evaluate its performance using a test set, computing the mean square error (MSE). Fifthly, adjust the orders while keeping the other structures of the model unchanged to train the model with new orders again, and select the orders of the model with the minimum MSE as the optimal orders. Next, keep the other structures of the model unchanged, adjust the network’s layer count, and select the network’s layer count with the smallest MSE as the optimal number of layers. Using the same method, determine the other model structures, i.e., the node count, activation function, and hyperparameters, and ultimately choose the model structure with the smallest MSE as the optimal model.

4. MPC Algorithm Design

This section designs the predictive controller based on models (6)–(8). By utilizing the model’s local linear and global nonlinear features, the predictive controller uses current state operating point information to locally linearize the model, calculates the nonlinear system’s predicted output, and obtains the control law by solving a quadratic programming (QP) problem.
For the convenience of MPC algorithm development, models (6)–(8) can be converted to the following forms:
{ y ( a ) = ρ 0 ( a 1 ) + q = 1 k g γ q , a 1 y ( a q ) + q = 1 k g λ q , a 1 u ( a q ) + θ ( a ) ρ 0 ( a 1 ) = ρ 0 ( ε ( a 1 ) ) k g = max ( s y , s u + s d 1 ) γ q , a 1 = { ρ y , q ( ε ( a 1 ) ) , ( q s y ) 0 , ( q > s y )   , λ q , a 1 = { ρ u , q ( ε ( a 1 ) ) , ( s d q s u + s d 1 ) 0 ,   else
where ρ 0 ( a 1 ) , γ q , a 1 , and λ q , a 1 indicate regression coefficients; k g represents the maximum value of s u + s d 1 and s y ; θ ( a ) denotes the white noise.
To transform the above equation into the state-space form, design the vectors below:
{ X ( a ) = [ x 1 , a T ,   x 2 , a T x k g , a T ] T , x 1 , a = y ( a ) x p , a = i = 1 k g + 1 p γ i + p 1 , a 1 y ( a i ) + j = 1 k g + 1 p λ j + p 1 , a 1 u ( a j ) p = 2 , 3 , , k g
Therefore, Equation (9) becomes the following form:
{ X ( a + 1 ) = G a X ( a ) + H a u ( a ) + ρ a + Λ ( a + 1 ) y ( a ) = Q X ( a )
where
{ G a = [ γ 1 , a I 0 0 γ 2 , a 0 I 0 γ k g 1 , a 0 0 I γ k g , a 0 0 0 ] k g × k g , H a = [ λ 1 , a   λ 2 , a     λ k g , a ] T , Q = [ I   0     0 ] ,   ρ a = [ ρ 0   0     0 ] T , Λ ( a + 1 ) = [ θ ( a + 1 )   0     0 ] T
Note that the coefficient matrix above is related to the working state ε ( a ) at time a . For different state-dependent ARX models, the values of matrices G a , H a , and ρ a in Equation (12) are different and can be calculated using Equation (9). To further develop the model predictive controller, define the vectors below:
{ X ( a ) = [ X ( a + 1 | a ) T X ( a + 2 | a ) T X ( a + M y | a ) T ] T Y ( a ) = [ y ( a + 1 | a ) T y ( a + 2 | a ) T y ( a + M y | a ) T ] T U ( a ) = [ u ( a ) T u ( a + 1 ) T u ( a + M u 1 ) T ] T ρ a = [ ρ a T ρ a + 1 T ρ a + M y 1 T ] T
where X ( a ) , Y ( a ) , U ( a ) , and ρ a represent the multi-step forward prediction state vector, multi-step forward prediction output vector, multi-step forward prediction control vector, and multi-step forward prediction offset, respectively; { X ( a + q | a ) T | q = 1 , , M y } and { y ( a + q | a ) T | q = 1 , , M y } represent the q -step forward prediction of the state and output; M y and M u represent the prediction and control time domains, respectively. The model’s multi-step forward prediction control is presented below:
{ X ( a ) = G a X ( a ) + H a U ( a ) + Ξ a ρ a , Y ( a ) = Q X ( a )
where
G a = [ i = 0 0 G a + i i = 0 1 G a + i i = 0 M y 1 G a + i ]     ,               i = j q G a + i = { G a + q G a + q 1 G a + j ,             j q I ,       j > q
Ξ a = [ I 0 0 0 i = 1 1 G a + i I 0 0 i = 1 2 G a + i i = 2 2 G a + i I 0 i = 1 M y 1 G a + i i = 2 M y 1 G a + i i = M y 1 M y 1 G a + i I ]    
H a = [ H a 0 0 ( i = 1 1 G a + i ) H a H a + 1 0 0 ( i = 1 M u 1 G a + i ) H a ( i = 2 M u 1 G a + i ) H a + 1 ( i = M u 1 M u 1 G a + i ) H a + M u 2 H a + M u 1 ( i = 1 M u G a + i ) H a ( i = 2 M u G a + i ) H a + 1 ( i = M u 1 M u G a + i ) H a + M u 2 j = M u 1 M u ( i = j + 1 M u G a + i ) H a + j ( i = 1 M y 1 G a + i ) H a ( i = 2 M y 1 G a + i ) H a + 1 ( i = M u 1 M y 1 G a + i ) H a + M u 2 j = M u 1 M y 1 ( i = j + 1 M y 1 G a + i ) H a + j ]
Q = [ Q 0 0 0 Q 0 0 0 Q ]
In Formula (14), the coefficient matrices G a , Ξ a , H a , and Q change with the system state and can be obtained by solving the system’s future working point ε ( a + q | a )   ( q = 1 , 2 , , M y 1 ) . Nevertheless, in the actual control process, it may be challenging to obtain information on the future state operating point. Therefore, compute using the current operating point ε ( a ) instead of ε ( a + q | a ) for the computer. Obtain the locally linearized model according to Equation (11) and use it to develop the following MPC method.
To facilitate the design of the objective function, Equation (14) is converted into the following forms:
{ Y ( a ) = R a U ( a ) + Y 0 ( a ) R a = Q H a Y 0 ( a ) = Q G a X ( a ) + Q Ξ a ρ a
Then, the desired output Y r ( a ) and control increment Δ U ( a ) are defined:
{ Δ U ( a ) = [ Δ u ( a ) T Δ u ( a + 1 ) T Δ u ( a + M u 1 ) T ] T Y r ( a ) = [ y r ( a + 1 ) T   y r ( a + 2 ) T y r ( a + M y ) T ] T
where Δ u ( a ) = u ( a ) u ( a 1 )   . The optimization objective function design for MPC as below:
{ min U ( a ) J = Y ( a ) Y r ( a ) A 1 2 + U ( a ) B 1 2 + Δ U ( a ) B 2 2 s . t . Y min Y ( a ) Y max , U min U ( a ) U max , Δ U min Δ U ( a ) Δ U max
where X Δ 2 = X T Δ X ; A 1 , B 1 , and B 2 represent weight coefficient matrices. Substitute Equations (19) and (20) into Equation (21) to eliminate the constant term and convert it into the following QP problem:
{ min U ( a ) J = 1 2 U ( a ) T [ R a T A 1 R a + B 1 + N T B 2 N 1 ] U ( a ) + [ Y 0 ( a ) T A 1 R a Y r ( a ) T A 1 R a U 0 ( a 1 ) T N T B 2 N 1 ] U ( a ) s . t . [ R a R a ] U ( a ) [ Y max Y 0 ( a ) Y min + Y 0 ( a ) ] , U min U ( a ) U max , U 0 ( a 1 ) + N Δ U min U ( a ) U 0 ( a 1 ) + N Δ U max .
where
{ U ( a ) = U 0 ( a 1 ) + N Δ U ( a ) , U 0 ( a 1 ) = [ U ( a 1 ) T U ( a 1 ) T U ( a 1 ) T ] T N = [ I 0 0 0 I I 0 0 I I I 0 I I I I ]
Equation (22) represents a convex QP problem with constraints. If the feasible set defined by the constraint condition is not empty and the objective function has a lower bound within the feasible set, then the QP problem has a globally optimal solution. If Equation (22) has no feasible solution in a certain control period, the feasible solution obtained in the previous period is used for control in the practical control. In addition, in practical control, in order to avoid control sequence deviations caused by environmental or system inherent biases, the optimal control sequence’s initial value is employed as the control input, and the system output is observed at the next moment for feedback correction, thus realizing online rolling optimization.

5. Control Experiments

This part uses the water tank system shown in Figure 6 as the experimental object, which is a commonly used process control device in industrial production. The deep learning models are established using the data-driven modeling approach, and the predictive controllers are designed based on the model’s distinctive structure for real-time control comparative experiments on water tank systems.
Figure 6. Water tank system [21].

5.1. Water Tank System

The water tank system is a representative dual-input, dual-output process control piece of experimental equipment with strong coupling, large time delay, and strong nonlinearity. The water inflows of water tanks 1 and 2 are controlled by electric control valves EV1 and EV2, the water outflows are regulated by proportional valves V1 and V2, and the water level heights are detected by the liquid level sensors LV1 and LV2. Control experiments were performed on the water tank system, and models (6)–(8) were designed and are shown below:
{ Y ( a ) = i = 1 s y G i , a 1 Y ( a i ) + j = s d s u + s d 1 H j , a 1 U ( a j ) + ρ 0 , a 1 + θ ( a ) G i , a 1 = [ γ i , a 1 11 γ i , a 1 12 γ i , a 1 21 γ i , a 1 22 ] , H j , a 1 = [ λ j , a 1 11 λ j , a 1 12 λ j , a 1 21 λ j , a 1 22 ] , ρ 0 , a 1 = [ ρ 0 , a 1 1 ρ 0 , a 1 2 ]
where Y ( a ) = [ y 1 ( a ) , y 2 ( a ) ] T represents the liquid level heights of tank 1 and tank 2; U ( a ) = [ u 1 ( a ) , u 2 ( a ) ] T represents the openings (0%~100%) of electric control valves EV1 and EV2; { G i , a 1 | i = 1 , , s y } , { H j , a 1 | j = s d , , s u + s d 1 } , and ρ 0 , a 1 are the weight coefficients, which can be obtained by models (6)–(8); θ ( a ) indicates the white noise signal. Among them, we have set the working point ε ( a 1 ) = [ Y ( a 1 ) T , , Y ( a d ) T ] T because the change in water level is an important factor leading to the strongly nonlinear dynamic characteristics of the water tank system.

5.2. Estimation of Model

In order to make the collected modeling data contain the various nonlinear dynamic characteristics of the water tank system to the maximum extent, this article uses the expected signal, including sine and step, to control the large fluctuation of the water tank level in the normal working range via the PID controller to obtain effective input/output sampling data. The time is 2 s. Figure 7 presents the structure of the PID controller used, which is mainly used to generate the system identification data. The PID controller output is limited between 0 and 100 (%).
Figure 7. The structure of MPC and PID control systems.
The sampling data given in Figure 8 are segregated into different proportions for model training and testing. The best proportion of the final modeling results is that the first 3500 data points (sampling time 2 s) are utilized for estimating the model, while the subsequent 1000 data points are utilized for verifying the model.
Figure 8. Identification data of water tank system.
The parameter optimization algorithms of the deep learning-based SD-ARX models are all the sampling adaptive motion estimation (Adam) algorithms, which integrate the advantages of momentum optimization and adaptive learning rate with high computational efficiency and can accelerate parameter convergence. The Tanh function is the full connecting layer activation function for all models, which gives the models a faster convergence rate and a stronger nonlinear expression ability. The hidden layer activation functions σ 4 ( ) and σ 5 ( ) of the LSTM are the Hard Sigmaid function and Tanh function, respectively. Compared with the standard Sigmoid function, the application of the Hard Sigmaid function in the neural network is more stable. It can reduce the gradient vanishing problem and has a faster calculation speed. The CNN employs the Relu as its convolutional activation function, which assists in reducing redundant calculation and parameter quantity between neurons, thus enhancing the model’s efficiency and generalization performance. The CNN pool function selects the average pool, which is beneficial for reducing the model’s complexity, minimizing the overfitting risk, and improving the feature’s robustness. Its convolutional and pooling layers have kernel sizes of 3 and 4. The MSE of different model structures is calculated and compared. When the model complexity meets the requirements, select the model structure and parameters with the lowest MSE for real-time control experiments. In addition, the search range for hyperparameters of all deep learning models includes the learning rate between 0.0001 and 0.1, the batch size between 4 and 128, and the epochs between 100 and 600. The hyperparameters finally selected include the Learning rate, batch size, and epochs of 0.001, 16, and 400, respectively. The parameter settings of the LSTM-CNN-ARX model proposed in this article are shown in Table 1.
Table 1. Parameter setting of the proposed model.
The model training in this study was implemented using the Keras [55] framework in the Jupyter Notebook 5.5.0 software development environment on a PC (Intel i9-11900k 3.5 GHz, 16 GB RAM). The supervisory control software used for the real-time control experiments was Kingview 6.55. Data communication between Matlab/Simulink 2022b software and Kingview 6.55 was achieved via the Dynamic Data Exchange (DDE) protocol, enabling real-time control of the water tank system. Taking the LSTM-CNN-ARX model as an example, Figure 9 and Figure 10 show that the model’s training and testing residuals are very small and the models have high modeling accuracy. Table 2 illustrates that different models have varying structures because they have distinct characteristics and nonlinear description capabilities. At the same time, the model with more parameters and deeper network layers has stronger nonlinear expression ability, but this also increases the computational burden, resulting in a longer training time. The findings suggest that the ARX model exhibits the maximum MSE and cannot capture the nonlinear system’s dynamic behavior well. The deep learning-based ARX model has a lower MSE than the RBF-ARX model because the multi-layer deep learning network has stronger nonlinear expression capabilities than the single-layer RBF neural network. The LSTM-ARX model has less MSE than the CNN-ARX model because LSTM can maintain long-term dependencies on sequence data. In addition, because the LSTM-CNN-ARX model incorporates the strengths of LSTM for learning temporal characteristics and CNN for capturing spatial characteristics, its MSE is smaller than the LSTM-ARX and CNN-ARX models, which can accurately capture the water tank system’s nonlinear behavior.
Figure 9. LSTM-CNN-ARX model’s training result.
Figure 10. LSTM-CNN-ARX model’s testing result.
Table 2. Comparison of modeling results.

5.3. Real-Time Control Experiments

This study was conducted in a laboratory environment using the water tank system experimental equipment shown in Figure 1, which uses the deep learning ARX model-based MPC method to compare with PID, ARX-MPC, and RBF-ARX-MPC methods, demonstrating the developed method’s superiority. The PID parameters were tuned for good performance at the given state of the water tank system. Figure 7 shows the control system’s structure, where the PID controller’s parameters are K P 1 = 18 ,   K I 1 = 1.2 ,   K D 1 = 0.2 and K P 2 = 20 ,   K I 2 = 1.1 ,   K D 2 = 0.2 . Because the water tank system operates in different liquid levels, it exhibits varying dynamic behaviors. Thus, control experiments are performed in different water level regions, i.e., low, medium, and high water level zones. The parameter setting of the objective optimization function (22) for the predictive control method is B 1 = diag [ 0.0001 ,   0.0001 ] , B 2 = diag [ 0.80 ,   0 . 80 ] , A 1 = diag [ 1 ,   1 ] . The control experimental results of the water tank system are shown in Figure 11, where y 1 ( a ) and y 2 ( a ) represent the liquid level heights, u 1 ( a ) and u 2 ( a ) represent the control valve opening (ranging from 0% to 100%), and y r ( a ) indicates that the expected output is a pink dashed line. In addition, Table 3, Table 4 and Table 5 display each algorithm’s control outcomes, encompassing overshoot (O), peak time (PT), and adjustment time (AT); and represent the rising and falling step response, respectively.
Figure 11. Low liquid level zone’s experimental results ( y 1 / u 1 ).
Table 3. Low liquid level zone’s control performance.
Table 4. Medium liquid level zone’s control performance.
Table 5. High liquid level zone’s control performance.

5.3.1. Low Liquid Level Zone Control Experiments

The control outcomes of distinct algorithms in a low liquid level zone (50–100 mm) are presented in Figure 11 and Figure 12 and Table 3, and the LSTM-CNN-ARX-MPC approach shows significantly superior control performance compared to other algorithms. In the falling step response of LSTM-CNN-ARX-MPC, the PTs of liquid levels y 1 ( a ) and y 2 ( a ) are 252 s and 224 s, the Os are 15.6% and 14.0%, and the ATs are 336 s and 320 s; in the rising step response, the Os are 3.2% and 3.6%, and the ATs are 166 s and 160 s, respectively, which are superior to other methods. This is due to the LSTM-CNN-ARX model incorporating the strengths of LSTM for learning temporal characteristics and CNN for capturing spatial characteristics, which can efficiently capture the water tank system’s nonlinear characteristics and achieve outstanding control performance. It should be noted that the PT of LSTM-CNN-ARX-MPC may not always outperform other methods. However, the evaluation of control performance is not just about looking at PT; we also need to see other indicators, such as O and AT. In fact, the O and AT of LSTM-CNN-ARX-MPC are much smaller than those of other methods, so overall, LSTM-CNN-ARX-MPC outperforms other methods. Usually, for controllers, their smaller PT may result in a larger O and AT, and there may also be significant oscillations at the beginning of the control process.
Figure 12. Low liquid level zone’s experimental results ( y 2 / u 2 ).

5.3.2. Medium Liquid Level Zone Control Experiments

Figure 13 and Figure 14 and Table 4 present the experimental results of various controllers in a medium liquid level zone (100–250 mm), indicating that the PID’s control performance is not as good as other methods; it has the maximum overshoot and AT. The control effectiveness of RBF-ARX-MPC outperforms ARX-MPC, exhibiting lower O and PT. However, RBF-ARX-MPC demonstrates inferior O and AT compared to the three deep learning-based MPC algorithms. Multi-layer deep learning networks have stronger nonlinear expression and generalization capabilities than single-layer RBF neural networks and can adaptively learn the water tank system’s nonlinear features. Therefore, the predictive control effect and modeling accuracy based on the deep learning ARX models are better than those of the RBF-ARX models. Additionally, LSTM-ARX-MPC’s comprehensive control effect is marginally superior to CNN-ARX-MPC because LSTM belongs to a recurrent neural network, which has storage function and advantages in time series data modeling, so it can maintain long-term correlation in sequence data. LSTM-CNN-ARX-MPC outperforms all other algorithms in the falling step response, exhibiting the lowest PT, AT, and O. Additionally, it has the smallest O in the rising step response, which enables the water tank system’s liquid level to effectively follow the reference trajectory.
Figure 13. Medium liquid level zone’s experimental results ( y 1 / u 1 ).
Figure 14. Medium liquid level zone’s experimental results ( y 2 / u 2 ).

5.3.3. High Liquid Level Zone Control Experiments

Figure 15 and Figure 16 and Table 5 present the experimental results in a high liquid level zone (250–300 mm). It is apparent that the PID controller exhibits the poorest control performance with the largest O and AT. ARX-MPC’s control performance is also poor because the ARX model has limited capability to capture nonlinear features. LSTM-ARX-MPC demonstrates superior overall control performance compared to ARX-MPC, RBF-ARX-MPC, and CNN-ARX-MPC due to the advantage of LSTM in maintaining long-term dependence on sequence data in time series modeling. LSTM-CNN-ARX-MPC demonstrates optimal comprehensive control effectiveness, particularly exhibiting a considerably lower O in the falling step. This is attributable to the LSTM-CNN-ARX model’s powerful spatiotemporal feature capture capability, which enable it to effectively capture the water tank system’s nonlinear properties.
Figure 15. High liquid level zone’s experimental results ( y 1 / u 1 ).
Figure 16. High liquid level zone’s experimental results ( y 2 / u 2 ).
In conclusion, PID and ARX-MPC exhibit the poorest control performance, failing to effectively capture the nonlinear characteristics of the water tank system, thereby limiting their applicability. Additionally, the overall control effectiveness of RBF-ARX-MPC is not as good as deep learning-based MPC. This is because RBF is a single-layer network with lesser nonlinear descriptive capabilities compared to multi-layer deep learning networks. Furthermore, LSTM-ARX-MPC’s superior control performance compared to CNN-ARX-MPC can be attributed to LSTM’s memory function, which preserves the long-term dependence on sequential data. In almost all cases, LSTM-CNN-ARX-MPC demonstrates the best control effectiveness, especially in the low and high water level regions with strong nonlinearity. This is attributed to the powerful spatiotemporal feature learning capability of the LSTM-CNN-ARX model, which enable it to capture all the nonlinear features of the water tank system.

6. Conclusions

This study aims to solve the modeling and predictive control problems of a category of smooth nonlinear systems. The studied system’s operating point may change over time, and its dynamic characteristics can be locally linearized. We utilized the spatiotemporal feature extraction capabilities of the LSTM and CNN, as well as the SD-ARX model’s pseudo-linear ARX structural feature, to establish the LSTM-CNN-ARX model to describe the studied system’s nonlinear properties. The developed modeling and MPC algorithms were validated in real-time control rather than digital simulation on an actual multivariable water tank system, confirming their effectiveness. The proposed methods can be useful and can generate better results for a category of smooth nonlinear plants than some well-established, simpler, but proven efficiency approaches.
This article used the deep learning model to fit the state-dependent ARX model’s autoregressive coefficients to obtain the LSTM-ARX model and the LSTM-CNN-ARX model that accurately describe the system’s nonlinear properties. These models have local linear and global nonlinear characteristics, which decompose the model’s complexity into the autoregressive part, making it easier to design the MPC controllers. Control comparative experiments demonstrate that compared with the PID, ARX-MPC, and RBF-ARX-MPC algorithms, the three deep learning-based MPC algorithms can achieve more precise trajectory tracking control of the water tank system’s liquid level. This is because of the stronger nonlinear expression and generalization ability of deep learning networks, which can adaptively learn system features and model parameters and better represent the nonlinear behavior of a water tank system. Furthermore, the comprehensive control performance of LSTM-CNN-ARX-MPC is superior to LSTM-ARX-MPC and CNN-ARX-MPC due to its incorporation of the strengths of LSTM for learning temporal characteristics and CNN for capturing spatial characteristics, which enables it to capture the multi-dimensional spatiotemporal nonlinear features of water tank systems more effectively.
While LSTM neural networks may effectively address the gradient vanishing or exploding problem in traditional RNN modeling and better handle long sequence data, the complex structure of the LSTM model with numerous unidentified parameters results in significant computational demands during model training, convergence difficulties, and the risk of overfitting. Consequently, future studies will concentrate on improving the model structure by replacing the LSTM structure with gated recurrent units or related improved structures to enhance modeling efficiency and expand the model’s applicability in scenarios with limited computational resources.

Author Contributions

Conceptualization, T.K. and H.P.; methodology, T.K.; software, T.K.; validation, H.P.; data curation, X.P.; writing—original draft preparation, T.K.; writing—review and editing, H.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Project of the National Natural Science Foundation of China (under Grant No. 61773402 and No. 52275105).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

All codes included in this study are available and can be obtained by contacting the author via email.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Mayne, D.Q. Model predictive control: Recent developments and future promise. Automatica 2014, 50, 2967–2986. [Google Scholar] [CrossRef]
  2. Elnawawi, S.; Siang, L.C.; O’Connor, D.L.; Gopaluni, R.B. Interactive visualization for diagnosis of industrial Model Predictive Controllers with steady-state optimizers. Control Eng. Pract. 2022, 121, 105056. [Google Scholar] [CrossRef]
  3. Kwapień, J.; Drożdż, S. Physical approach to complex systems. Phys. Rep. 2012, 515, 115–226. [Google Scholar] [CrossRef]
  4. Chen, H.; Jiang, B.; Ding, S.X.; Huang, B. Data-driven fault diagnosis for traction systems in high-speed trains: A survey, challenges, and perspectives. IEEE Trans. Intell. Transp. Syst. 2022, 23, 1700–1716. [Google Scholar] [CrossRef]
  5. Yu, W.; Wu, M.; Huang, B.; Lu, C. A generalized probabilistic monitoring model with both random and sequential data. Automatica 2022, 144, 110468. [Google Scholar] [CrossRef]
  6. Pareek, P.; Verma, A. Piecewise Linearization of Quadratic Branch Flow Limits by Irregular Polygon. IEEE Trans. Power Syst. 2018, 33, 7301–7304. [Google Scholar] [CrossRef]
  7. Ławryńczuk, M. Nonlinear predictive control of a boiler-turbine unit: A state-space approach with successive on-line model linearisation and quadratic optimisation. ISA Trans. 2017, 67, 476–495. [Google Scholar] [CrossRef] [PubMed]
  8. Qian, K.; Zhang, Y. Bilinear model predictive control of plasma keyhole pipe welding process. J. Manuf. Sci. Eng. 2014, 136, 31002. [Google Scholar] [CrossRef]
  9. Nie, Z.; Gao, F.; Yan, C.B. A multi-timescale bilinear model for optimization and control of HVAC systems with consistency. Energies 2021, 14, 400. [Google Scholar] [CrossRef]
  10. Shi, Y.; Yu, D.L.; Tian, Y.; Shi, Y. Air-fuel ratio prediction and NMPC for SI engines with modified Volterra model and RBF network. Eng. Appl. Artif. Intell. 2015, 45, 313–324. [Google Scholar] [CrossRef]
  11. Gruber, J.K.; Ramirez, D.R.; Limon, D.; Alamo, T. A convex approach for NMPC based on second order Volterra series models. Int. J. Robust Nonlinear Control 2015, 25, 3546–3571. [Google Scholar] [CrossRef]
  12. Chen, H.; Chen, Z.; Chai, Z.; Jiang, B.; Huang, B. A single-side neural network-aided canonical correlation analysis with applications to fault diagnosis. IEEE Trans. Cybern. 2021, 52, 9454–9466. [Google Scholar] [CrossRef]
  13. Chen, H.; Liu, Z.; Alippi, C.; Huang, B.; Liu, D. Explainable intelligent fault diagnosis for nonlinear dynamic systems: From unsupervised to supervised learning. IEEE Trans. Neural Netw. Learn. Syst. 2022. [Google Scholar] [CrossRef]
  14. Ding, B.; Wang, J.; Su, B. Output feedback model predictive control for Hammerstein model with bounded disturbance. IET Control. Theory Appl. 2022, 16, 1032–1041. [Google Scholar] [CrossRef]
  15. Raninga, D.; TK, R.; Velswamy, K. Explicit nonlinear predictive control algorithms for Laguerre filter and sparse least square support vector machine-based Wiener model. Trans. Inst. Meas. Control 2021, 43, 812–831. [Google Scholar] [CrossRef]
  16. Wang, Z.; Georgakis, C. Identification of Hammerstein-Weiner models for nonlinear MPC from infrequent measurements in batch processes. J. Process Control 2019, 82, 58–69. [Google Scholar] [CrossRef]
  17. Du, J.; Zhang, L.; Chen, J.; Li, J.; Zhu, C. Multi-model predictive control of Hammerstein-Wiener systems based on balanced multi-model partition. Math. Comput. Model. Dyn. Syst. 2019, 25, 333–353. [Google Scholar] [CrossRef]
  18. Peng, H.; Ozaki, T.; Haggan-Ozaki, V.; Toyoda, Y. A parameter optimization method for radial basis function type models. IEEE Trans. Neural Netw. 2003, 14, 432–438. [Google Scholar] [CrossRef]
  19. Zhou, F.; Peng, H.; Qin, Y.; Zeng, X.; Xie, W.; Wu, J. RBF-ARX model-based MPC strategies with application to a water tank system. J. Process Contr. 2015, 34, 97–116. [Google Scholar] [CrossRef]
  20. Peng, H.; Nakano, K.; Shioya, H. Nonlinear predictive control using neural nets-based local linearization ARX model—Stability and industrial application. IEEE Trans. Control Syst. Technol. 2006, 15, 130–143. [Google Scholar] [CrossRef]
  21. Kang, T.; Peng, H.; Zhou, F.; Tian, X.; Peng, X. Robust predictive control of coupled water tank plant. Appl. Intell. 2021, 51, 5726–5744. [Google Scholar] [CrossRef]
  22. Peng, H.; Ozaki, T.; Toyoda, Y.; Shioya, H.; Nakano, K.; Haggan-Ozaki, V.; Mori, M. RBF-ARX model-based nonlinear system modeling and predictive control with application to a NOx decomposition process. Control Eng. Pract. 2004, 12, 191–203. [Google Scholar] [CrossRef]
  23. Peng, H.; Wu, J.; Inoussa, G.; Deng, Q.; Nakano, K. Nonlinear system modeling and predictive control using the RBF nets-based quasi-linear ARX model. Control Eng. Pract. 2009, 17, 59–66. [Google Scholar] [CrossRef]
  24. Xu, W.; Peng, H.; Tian, X.; Peng, X. DBN based SD-ARX model for nonlinear time series prediction and analysis. Appl. Intell. 2020, 50, 4586–4601. [Google Scholar] [CrossRef]
  25. Inoussa, G.; Peng, H.; Wu, J. Nonlinear time series modeling and prediction using functional weights wavelet neural network-based state-dependent AR model. Neurocomputing 2012, 86, 59–74. [Google Scholar] [CrossRef]
  26. Mu, R.; Zeng, X. A review of deep learning research. KSII Trans. Internet Inf. Syst. 2019, 13, 1738–1764. [Google Scholar] [CrossRef]
  27. Lippi, M.; Montemurro, M.A.; Degli Esposti, M.; Cristadoro, G. Natural language statistical features of LSTM-generated texts. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 3326–3337. [Google Scholar] [CrossRef]
  28. Shuang, K.; Tan, Y.; Cai, Z.; Sun, Y. Natural language modeling with syntactic structure dependency. Inf. Sci. 2020, 523, 220–233. [Google Scholar] [CrossRef]
  29. Shuang, K.; Li, R.; Gu, M.; Loo, J.; Su, S. Major-minor long short-term memory for word-level language model. IEEE Trans. Neural Netw. Learn. Syst. 2019, 31, 3932–3946. [Google Scholar] [CrossRef]
  30. Ying, W.; Zhang, L.; Deng, H. Sichuan dialect speech recognition with deep LSTM network. Front. Comput. Sci. 2020, 14, 378–387. [Google Scholar] [CrossRef]
  31. Jo, J.; Kung, J.; Lee, Y. Approximate LSTM computing for energy-efficient speech recognition. Electronics 2020, 9, 2004. [Google Scholar] [CrossRef]
  32. Oruh, J.; Viriri, S.; Adegun, A. Long short-term Memory Recurrent neural network for Automatic speech recognition. IEEE Access 2022, 10, 30069–30079. [Google Scholar] [CrossRef]
  33. Yang, B.; Yin, K.; Lacasse, S.; Liu, Z. Time series analysis and long short-term memory neural network to predict landslide displacement. Landslides 2019, 16, 677–694. [Google Scholar] [CrossRef]
  34. Hu, J.; Wang, X.; Zhang, Y.; Zhang, D.; Zhang, M.; Xue, J. Time series prediction method based on variant LSTM recurrent neural network. Neural Process. Lett. 2020, 52, 1485–1500. [Google Scholar] [CrossRef]
  35. Peng, L.; Zhu, Q.; Lv, S.X.; Wang, L. Effective long short-term memory with fruit fly optimization algorithm for time series forecasting. Soft Comput. 2020, 24, 15059–15079. [Google Scholar] [CrossRef]
  36. Langeroudi, M.K.; Yamaghani, M.R.; Khodaparast, S. FD-LSTM: A Fuzzy LSTM Model for Chaotic Time-Series Prediction. IEEE Intell. Syst. 2022, 37, 70–78. [Google Scholar] [CrossRef]
  37. Wu, Z.; Rincon, D.; Luo, J.; Christofides, P.D. Machine learning modeling and predictive control of nonlinear processes using noisy data. AICHE J. 2021, 67, e17164. [Google Scholar] [CrossRef]
  38. Terzi, E.; Bonassi, F.; Farina, M.; Scattolini, R. Learning model predictive control with long short-term memory networks. Int. J. Robust Nonlinear Control 2021, 31, 8877–8896. [Google Scholar] [CrossRef]
  39. Zarzycki, K.; Ławryńczuk, M. Advanced predictive control for GRU and LSTM networks. Inf. Sci. 2022, 616, 229–254. [Google Scholar] [CrossRef]
  40. Yu, W.; Zhao, C.; Huang, B. MoniNet with concurrent analytics of temporal and spatial information for fault detection in industrial processes. IEEE Trans. Cybern. 2021, 52, 8340–8351. [Google Scholar] [CrossRef]
  41. Li, G.; Huang, Y.; Chen, Z.; Chesser, G.D.; Purswell, J.L.; Linhoss, J.; Zhao, Y. Practices and applications of convolutional neural network-based computer vision systems in animal farming: A review. Sensors 2021, 21, 1492. [Google Scholar] [CrossRef] [PubMed]
  42. Yang, X.; Ye, Y.; Li, X.; Lau, R.Y.; Zhang, X.; Huang, X. Hyperspectral image classification with deep learning models. IEEE Trans. Geosci. Remote Sens. 2018, 56, 5408–5423. [Google Scholar] [CrossRef]
  43. Gu, J.; Wang, Z.; Kuen, J.; Ma, L.; Shahroudy, A.; Shuai, B.; Chen, T. Recent advances in convolutional neural networks. Pattern Recognit. 2018, 77, 354–377. [Google Scholar] [CrossRef]
  44. Jiang, M.; Zhang, W.; Zhang, M.; Wu, J.; Wen, T. An LSTM-CNN attention approach for aspect-level sentiment classification. J. Comput. Methods Sci. Eng. 2019, 19, 859–868. [Google Scholar] [CrossRef]
  45. Kumar, B.S.; Malarvizhi, N. Bi-directional LSTM-CNN combined method for sentiment analysis in part of speech tagging (PoS). Int. J. Speech Technol. 2020, 23, 373–380. [Google Scholar] [CrossRef]
  46. Priyadarshini, I.; Cotton, C. A novel LSTM-CNN-grid search-based deep neural network for sentiment analysis. J. Supercomput. 2021, 77, 13911–13932. [Google Scholar] [CrossRef]
  47. Jang, B.; Kim, M.; Harerimana, G.; Kang, S.U.; Kim, J.W. Bi-LSTM model to increase accuracy in text classification: Combining Word2vec CNN and attention mechanism. Appl. Sci. 2020, 10, 5841. [Google Scholar] [CrossRef]
  48. Shrivastava, G.K.; Pateriya, R.K.; Kaushik, P. An efficient focused crawler using LSTM-CNN based deep learning. Int. J. Syst. Assur. Eng. Manag. 2022, 14, 391–407. [Google Scholar] [CrossRef]
  49. Zhu, Y.; Gao, X.; Zhang, W.; Liu, S.; Zhang, Y. A bi-directional LSTM-CNN model with attention for aspect-level text classification. Future Internet 2018, 10, 116. [Google Scholar] [CrossRef]
  50. Zhen, H.; Niu, D.; Yu, M.; Wang, K.; Liang, Y.; Xu, X. A hybrid deep learning model and comparison for wind power forecasting considering temporal-spatial feature extraction. Sustainability 2020, 12, 9490. [Google Scholar] [CrossRef]
  51. Farsi, B.; Amayri, M.; Bouguila, N.; Eicker, U. On short-term load forecasting using machine learning techniques and a novel parallel deep LSTM-CNN approach. IEEE Access 2021, 9, 31191–31212. [Google Scholar] [CrossRef]
  52. Hoiberg, J.A.; Lyche, B.C.; Foss, A.S. Experimental evaluation of dynamic models for a fixed-bed catalytic reactor. AICHE J. 1971, 17, 1434–1447. [Google Scholar] [CrossRef]
  53. O’Hagan, A. Curve fitting and optimal design for prediction. J. R. Stat. Soc. Ser. B Methodol. 1978, 40, 1–24. [Google Scholar] [CrossRef]
  54. Priestley, M.B. State-dependent models: A general approach to non-linear time series analysis. J. Time Ser. Anal. 1980, 1, 47–71. [Google Scholar] [CrossRef]
  55. Ketkar, N. Introduction to Keras. In Deep Learning with Python; Apress: Berkeley, CA, USA, 2017; pp. 97–111. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.