Next Article in Journal
Combining Distributed Consensus with Robust H-Control for Satellite Formation Flying
Previous Article in Journal
Network-Oriented Real-Time Embedded System Considering Synchronous Joint Space Motion for an Omnidirectional Mobile Robot
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Design of Neural Network Quantizers for Networked Control Systems

by
Juan Esteban Rodriguez Ramirez
1 and
Yuki Minami
2,*
1
Graduate School of Information Science, Nara Institute of Science and Technology, 8916-5 Takayamacho, Ikoma, Nara 630-0192, Japan
2
Department of Mechanical Engineering, Graduate School of Engineering, Osaka University, 2-1 Yamadaoka, Suita, Osaka 565-0871, Japan
*
Author to whom correspondence should be addressed.
Electronics 2019, 8(3), 318; https://doi.org/10.3390/electronics8030318
Submission received: 24 February 2019 / Revised: 6 March 2019 / Accepted: 7 March 2019 / Published: 13 March 2019
(This article belongs to the Section Networks)

Abstract

:
Nowadays, networked control systems (NCSs) are being widely implemented in many applications. However, several problems negatively affect and compromise the design of practical NCSs. One of them is the performance degradation of the system due to quantization. This paper aims to develop dynamic quantizers for NCSs and their design methods that alleviate the effects of the quantization problem. In this paper, we propose a type of dynamic quantizers implemented with neural networks and memories, which can be tuned by a time series data of the plant inputs and outputs. Since the proposed quantizer can be designed without the model information of the system, the quantizer could be applied to any system with uncertainty or nonlinearity. This paper gives two types of quantizers, and they differ from each other in the neural networks structure. The effectiveness of these quantizers and their design method are verified using numerical examples. Besides, their performances are compared among each other using statistical analysis tools.

1. Introduction

The networked control systems (NCSs) are systems in which its elements are physically separated but connected by some communication channels. They have been around for some decades already, and they have been implemented successfully in many fields such as industrial automation, robotics, and power grids. Although the NCSs provide several advantages to the systems, it is well-known that one of the problems is the system performance degradation caused by the data rate constraints in the communication channels [1,2,3]. In the case that operation signals of NCSs are transmitted over networks under data rate constraints, the signal quantization is a fundamental process in which a continuous-valued signal is transformed into a discrete-valued one. However, the quantization error between the continuous-valued signal and discrete-valued one occurs, and it affects the performance of the NCSs. Therefore, one of the significant works is to develop a method to minimize the influence of the quantization error to the performance of the NCSs.
It has been proven that properly designed feedback-type dynamic quantizers are effective to reduce this degradation in the system’s performance [4]. Several studies have considered the design of dynamic quantizers. For instance, a mathematical expression of an optimal dynamic quantizer for time-invariant linear plants was presented in [4], and an equivalent expression for nonlinear plants was introduced in [5]. Furthermore, design methods for dynamic quantizers that minimize the system’s performance degradation and satisfy the channel’s data rate constraints were developed in [6,7,8]. Then, in [9] event-triggered dynamic quantizers that reduce the traffic in the communication network were proposed. In these studies, the design of the quantizers is carried out using information from the plant; namely, these quantizers are based on model-based approach. Thus, if the model of the plant is inaccurate, then the quantizers will be faulty.
Accordingly, in this paper, the data-driven approach is considered for the design of feedback type dynamic quantizers. Besides, this paper presents a class of dynamic quantizers that are constructed using feedforward neural networks. The quantizer, called neural network quantizers, are designed using time series data of plant inputs and outputs. Some advantages of this approach are that a model of the plant is not required for the design, i.e., model-free design, and that the quantizer can be designed not only for linear but also for nonlinear plants. The selection of neural networks to perform this job is motivated by the fact that feedforward neural networks are very flexible and that they can be used to represent any nonlinear function/system, in this sense, they work as universal approximators [10,11]. This property is especially important for the design of optimal dynamic quantizers because their structures are functions of the plants’ model [4,5]. If the model of the plant is not given, but it is known to be linear, then the structure of the optimal quantizer is also known to be linear. However, if the plant is nonlinear and its model is absent, then the optimal quantizer’s structure is unknown. Thus, the neural network can approximate the optimal quantizer’s structure based on a series of plant input and output data.
This paper is structured as follows. First, we propose a class of dynamic quantizers composed of feedforward neural network, memories, and a static quantizer. The proposed quantizer has two variations in neural network structures: one is based on a regression-based approach, and the other is based on a classification-based approach. Then, we formulate the quantizers design problem that finds the parameters of the neural network and the quantization interval for given a time series data of plant input and output. Next, with numerical examples, the effectiveness of these quantizers and their design method are verified. Finally, several design variations are considered in order to optimize the quantizer’s performance, and comparisons among these variations are carried out.
It should be remarked that various results on the quantizer design for networked control systems have been obtained, e.g., [12,13,14,15,16]. However, the contributions of this paper are distinguished from them as follows. The papers in [12,13,14] focus on the zoom-in and zoom-out strategy based dynamic quantizer, i.e., the quantizer with time-varying quantization interval. Besides, the paper [15] considers the static logarithmic quantizer, i.e., its quantization interval is not uniform. On the other hand, this paper proposes dynamic quantizer with the time-invariant and uniform interval. Furthermore, the paper [16] proposes a ΔΣ modulator, which is related to the proposed quantizer. Although the result in [16] is restricted to the case of two quantization levels, this paper can deal with the case of multi-levels.
This paper is a journal version of our previous conference papers that were presented in [17,18]. The main difference between this paper and its predecessors is as follows. The system description and problem formulation are improved, and detail explanations of the proposed quantizers are added. Then, we use the ANOVA test to analyze the several simulation results. Besides, we take into account different activation functions for the neural networks hidden layers and compares different initialization methods for the network tuning.

2. Neural Network Quantizers

In this section, we first describe a networked control system considered here. Then we present a quantizer, composed of neural networks and a static quantizer, called neural network quantizer.

2.1. System Description

This paper considers the system depicted in Figure 1. This system is composed of a plant P, a communication channel that has no loses or delays, and the neural network quantizer Q N N proposed in this paper.
The plant is represented by the following single-input-single-output (SISO) model:
P : x ( k + 1 ) = f x ( k ) , v ( k ) , y ( k ) = g x ( k ) ,
where k { 0 } N is the discrete time, x R n P is the state vector with initial value x ( 0 ) = x 0 , v R is the input, and y R is the output. The functions f : R n P × R R n P and g : R n P R are in general nonlinear mappings. It is assumed that f and g are continuous and smooth.
The quantizer Q N N , shown in Figure 1, is composed of a neural network, a static quantizer q , and a couple of memories S v and S y . The quantizer is represented are by the following expression:
Q N N : y ( k ) = Γ 1 u ( k ) , V ( k ) , Y q ( k ) , u q ( k ) = Γ 2 y ( k ) , v ( k ) = q u q ( k ) ,
where u R is the input, y R K n L is the output of the neural network, u q R is the input of q , and v { ± d 2 , ± 2 d 2 , , ± M 2 d 2 } is the output of q , i.e., the output of Q N N . Note that d is the quantization interval, and M is the number of quantization levels which is determined from the data rate of network channel. The signals V ( k ) and Y q ( k ) are the outputs of the memories S v and S y , respectively. They are time series of past values of the quantized inputs v ( k ) and the outputs y q ( k ) of the plant, and they are given by
V ( k ) = [ v ( k 1 ) , v ( k 2 ) , , v ( k n V ) ] ,
Y q ( k ) = [ y q ( k 1 ) , y q ( k 2 ) , , y q ( k n Y ) ] ,
where n V and n Y are the dimensions of these memories. Thus, the proposed quantizer is tuned by using the past input and output data of the plant. This means that the quantizer may capture the dynamics of the plant.
This paper proposes two types of neural network quantizers: Q N N R and Q N N C . The quantizers Q N N R and Q N N C differ in the expressions of the nonlinear functions Γ 1 ( · ) , Γ 2 ( · ) , and q ( · ) . The illustrations of Q N N R and Q N N C are shown in Figure 2. Although detail explanations of two quantizers will be shown in the following subsections, the main difference between them is in the neural network’s structure. In Q N N R the network has only one output that shapes the input signal, i.e., the network is trained to perform regression. On the other hand, in Q N N C the network has as many outputs as the considered amount of quantization levels M. Each output represents the probability that a given input is matched with a specific quantization level, i.e., the network is trained for classification. Besides, in Figure 2a, the numbers 2 , 1 , 1 , 3 , 3 mean the selected quantization levels, which are determined by the function Γ 2 .

2.2. Regression Based Neural Network Quantizer

For the regression-based neural network quantizer Q N N R , the static quantizer q ( · ) is a regular finite-level static quantizer with saturation as shown in Figure 3. It receives directly the continuous output of the neural network u q ( k ) and rounds it to the nearest discrete value to generate v ( k ) . It has two parameters: one is the number of quantization levels M N and the other is the quantization interval d R with d > 0 . Figure 3 shows an example of this static quantizer with M = 4 .
In this paper, the fully connected feed-forward type neural network is adopted to build the function Γ 1 ( · ) in Equation (2). An example is shown in Figure 4. In this case, the function Γ 1 is given by
Γ 1 : y l = k = 0 K 2 w k l ( 3 ) h j = 0 K 1 w k j ( 2 ) h i = 0 K 0 w j i ( 1 ) x i for l = 1 , 2 , , K n L .
The following elements can be recognized in the network, the input units x R K 0 , the output units y R K n L , and the hidden units z ( i ) R K i ( i = 1 , 2 , , n L 1 ), where n L is the number of layers in the network. Note that the inputs of this network are u ( k ) , V ( k ) and Y q ( k ) , and they are represented as x ( k ) = [ u ( k ) , V ( k ) , Y q ( k ) ] . Besides, the size of the neural network is represented by K = K 1 , K 2 , , K n L . Each neuron performs a nonlinear transformation of a weighted summation of the previous layer outputs as follows
z j ( l ) = h i = 0 K l w j i ( l ) z i ( l 1 ) ,
where w j i ( l ) represents the weight of the connection that goes from the ith neuron in layer ( l 1 ) to the jth neuron in layer l. Notice here that a simplified notation is used, where instead of having biases the units x 0 = 1 and z 0 ( l 1 ) = 1 are included in the network. Then, because these elements are constants, their respective connection’s weight w j 0 ( l ) serves as bias parameters. The weights of all the connections in the network are put together in a vector w called the weights vector that has dimension n w = i = 0 n L 1 ( K i + 1 ) K i + 1 . Furthermore, h ( · ) represents the nonlinear transformation and is called activation function. There are many functions that serve as activation functions such as logistic sigmoid, hyperbolic tangent, and rectified linear unit (ReLU). In this paper, we adopt the most commonly used sigmoid function:
sigm ( a ) = 1 1 + exp ( a ) .
Finally, since K n L = 1 in Q N N R , the function Γ 2 is given by Γ 2 : u q = y l .

2.3. Classification Based Neural Network Quantizer

For the classification based neural network quantizer Q N N C , the static quantizer q ( · ) is not a conventional one. Its input u q ( k ) comes from a set of indexes, each of which makes reference to a specific quantization level, i.e., u q ( k ) { 1 , 2 , , M } . Thus, q ( · ) is adapted to match each index to the corresponding quantization level as Figure 5 shows. This quantizer is also defined by the number of quantization levels M and the quantization interval d.
The neural network Γ 1 in Q N N C is same as that in Q N N R . The inputs of this network are the same as in the previous case x ( k ) = [ u ( k ) , V ( k ) , Y q ( k ) ] and the hidden units activation function h ( · ) is also the logistic sigmoid in Equation (7). The dimension of the ouput is K n L = M . Then each output of the network y i ( k ) is associated with one quantization level, and represents the probability that a given input is classified into a specific quantization level. Therefore, the quantization level with the biggest probability is selected to be the network’s output, and it is given by
Γ 2 : u q ( k ) = arg max i { 1 , 2 , , M } exp ( y i ( k ) ) j = 1 K n L exp ( y j ( k ) ) .

3. Quantizer Design Problem

In this paper, it is assumed that the number of quantization levels M, the memory sizes n V and n Y , and the neural network structure K are given. Thus, the design parameters are the weight vector w and the quantization interval d of the neural network quantizer Q N N .
The performance of the quantizer Q N N in NCSs can be evaluated using a construction known as error system. The considered error system is depicted in Figure 6. This system is composed of two branches. In the lower branch, the input signal u is applied directly to the plant P that produces the ideal output y. In the upper branch the effects of quantization are considered and u is applied to the quantizer Q N N that generates the quantized signal v that is applied to the plant. The output of the plant in this case is represented by y q , and the difference y q y is the error signal. The error signal e ( k ) is used to evaluate the performance degradation of the system. By minimizing y q y , the system composed of the quantizer Q N N and the plant P can be optimally approximated to the plant P, in terms of the input-output relation.
In this context, a parameter known as performance index is used to measure the system’s performance degradation. The performance index considered here is the sum-of-squares error function that is defined by
E ( Q N N ) = k = 0 n s y q ( u ( k ) , w , d ) y ( k ) 2
where u ( k ) is used to build x ( k ) along side with V ( k ) and Y q ( k ) that are generated dynamically. It is necessary to make E ( Q N N ) as small as possible to maintain the output error low. Then, the design of Q N N is set up as an optimization problem in which the performance index is minimized.
This paper assumes that, although the model is unknown, it is possible to feed it with inputs and measure its outputs. Then, a time series of inputs and outputs of the plant will be available. These time series are represented as follows.
U = [ u ( 1 ) , u ( 2 ) , , u ( n s ) ] ,
Y = [ y ( 1 ) , y ( 2 ) , , y ( n s ) ] ,
where n s is the length of the time series, namely, the number of samples. Notice that y ( k ) ( k = 1 , 2 , , n s ) represents the output of the plant P when u ( k ) is applied directly to it, i.e., v ( k ) = u ( k ) .
Then, the neural network quantizers design problem is formulated as follows:
Problem 1.
Suppose that the time series data U and Y of the plant, the number of quantization levels M, the neural network structure K , and the memory sizes n V and n Y are given. Then, find the parameters of Q N N : the weight vector w and the quantization interval d which minimize E ( Q N N ) , under the condition that d > 0 .
This design problem is nonlinear and nonconvex. Thus, it cannot be solved using gradient-based optimization methods such as linear programming or quadratic programming. Moreover, conventional neural network training techniques based on error backpropagation cannot be used either due to the structure of the system, as it was explained previously. Therefore, alternative optimization methods should be used.
In this regard, the metaheuristics stand out from the available options because of their flexibility and a wide variety of implementations [19]. In particular, the differential evolution (DE) metaheuristic algorithm is used to perform the design of Q N N . This choice is justified by the fact that DE has proven to be effective in the training of neural networks [20,21] and that it has shown an outstanding performance in the design of dynamic quantizers [9]. DE is a population-based metaheuristic algorithm inspired in the mechanism of biological evolution [22,23]. In this algorithm, the cost function J ( θ ) is evaluated iteratively over a population of possible solutions or individuals θ i R n in each iteration the individuals improve their values and move towards the best solution. Finally, the individual with the lowest fitness value in the last iteration is regarded as the optimal solution. Some advantages of DE are that it is very easy to implement and has only two tuning parameters: the scale factor F and the crossover constant H, apart from the number of individuals N and the maximum number of iterations t m a x . Besides, DE shows very good exploration capacities and converges fast to global optima. DE has many versions and variations; the one considered in this study is the classical DE/best/1/bin strategy, which is described in Algorithm 1.
Algorithm 1: DE (DE/best/1/bin strategy)
Initialization: Given N N , t m a x N , F [ 0 , 2 ] , H [ 0 , 1 ] and the initial search space S 0 = [ θ m i n , θ m a x ] n . Set t = 0 then select randomly N individuals { θ 1 , θ 2 , , θ N } in the search space.
Step 1: The cost function J ( θ ) is evaluated for each θ i and θ b a s e = θ l is calculated by:
l = arg min i { 1 , 2 , , N } J ( θ i ) .
If t = t m a x then θ b a s e is the final solution, if not go to Step 2.
Step 2 (Mutation): For each θ i a mutant vector M i is generated by:
M i = θ b a s e + F ( θ τ 1 , i θ τ 2 , i ) ,
where τ 1 , i and τ 2 , i are random indexes subject to i τ 1 , i τ 2 , i l .
Step 3 (Crossover): For each θ i and M i a trial vector T i is generated by:
T i , j = M i , j if ρ i , j H or j = j r a n d , θ i , j otherwise ,
where ρ i , j [ 0 , 1 ] and j r a n d { 1 , 2 , , n } are generated randomly.
Step 4 (Selection): The members of the next generation k + 1 are selected by:
θ i T i if J ( T i ) J ( θ i ) , θ i otherwise ,
then t t + 1 and go to Step 1.
Since the design parameters of Q N N are w and d, an individual for the DE algorithm will have the following form θ = [ d w ] with dimension n = 1 + n w . From these parameters, the weights vector w is not affected by any constraint, but the quantization interval d should always be positive d > 0 . DE has no direct way to handle the constraints of the optimization problem since it was designed to solve unconstrained problems. Then, in order to manage the constraint condition, a method developed by Maruta et al. in [24] is employed. This method transforms the constrained optimization problem into the following unconstrained one.
minimize θ R n J ( θ ) for J ( θ ) : = arctan [ E ( θ ) ] π / 2 if d > 0 , d otherwise ,
where E ( θ ) is the performance index in Equation (9). This constraints management method ensures that d is positive.
The learning resulting from the training of a deep neural network depends highly on the initial weights of the network because many of the learning techniques are in essence local searches. Therefore, it is very important to initialize the network’s weights appropriately [25,26]. There are several ways to initialize the neural networks to perform the training. The most common method is the uniformly random initialization where random values sampled from a certain interval using a uniform probability distribution are assigned to the weights and biases of the network. The initialization intervals are selected according to, but they are usually small and close to zero. Popular ones are the intervals [ 1 , 1 ] or [ 0.5 , 0.5 ] . Another prevalent type of initialization was developed in [27] by Glorot and Bengio. This method is known as Xavier Uniform initialization (from Xavier Glorot). In this method, the weights of each layer in the network are initialized using random uniform sampling in a specific interval
w i U [ l i , l i ] for i = 1 , 2 , , n L ,
where w i represents the weights of the ith layer. The limits of the interval are given by l i which is a function of the number of neurons of the considered layer K i , the number of neurons in the previous layer K i 1 and the hidden layers activation function h . The limit is the following
l i = 4 6 K i 1 + K i for h ( a ) = sigm ( a ) .

4. Numerical Simulations

To verify that the proposed neural network quantizers and their design method work properly, several numerical simulations were performed. In these simulations, the following discrete, nonlinear and stable plant is used.
P : x 1 ( k + 1 ) x 2 ( k + 1 ) = f 1 ( x ( k ) ) + f 3 ( x ( k ) ) u ( k ) f 2 ( x ( k ) ) + f 3 ( x ( k ) ) u ( k ) , y ( t ) = 1.45 x 1 ( k ) + x 2 ( k ) ,
f 1 ( x ) = 0.8 x 1 0.4 x 2 + 0.4 e | x 2 | cos 3 ( x 1 ) , f 2 ( x ) = 0.6 x 2 + 0.4 e | x 1 | cos ( x 1 ) 1 2 , f 3 ( x ) = 0.01 + 0.01 ( ( x 1 ) 4 + 0.1 ) 1 .
This plant is a modified version of the plant shown in [5]. The initial state is x 0 = [ 0.1 , 0.2 ] , and the input signal used in the examples is given by
u ( k ) = 0.3 sin ( 6 k ) + 0.4 sin ( k ) + 0.3 sin ( 3 k ) .
The evaluation interval is L = 1000 , which implies that the amount of samples taken is n s = 1000 .
The quantizers are constructed with n Y = n V = 5 , M = { 2 , 8 } and neural networks with n L = { 2 , 4 } . Given the size of the memories and the dimension of u ( k ) all the networks have inputs with dimension K 0 = 11 . The neural networks’ structure depends on the type of quantizer and M. Table 1 summaries the structure of the quantizers used in the simulations. For the regression case (R) the network’s structure and the dimension of w ( n w ) are independent of M. This is not the case for the classification type of quantizer (C). Table 1 also shows a comparison among the n w of each network.
The hyper parameters of DE are N = 500 , t m a x = 2000 , F = 0.6 , and H = 0.9 . The simulations were performed N r u n = 50 times for each considered case. Then, since the individuals have the form θ = [ d w ] the dimensions of the optimization problems n will be the ones shown is the last column of Table 1. Looking at Table 1 it is possible to see that Q N N C has more parameters than Q N N R , this is a factor that influences the performance of the proposed design method.
The DE individuals are initialized as follows. The first element d is uniformly sampled from the interval ( 0 , 1 ] . The network weights are initialized using the uniform random and the Xavier uniform initialization methods, described in Section 3.
After running the DE algorithm N r u n = 50 times for each considered case, the quantizers Q N N with the lowest E ( w , d ) are selected to be the optimal quantizers. Then, in order to test these quantizers, the error system in Figure 6 is fed with the input signal u ( k ) for each case. It results that all the quantizers work properly and show good performance. For instance, Figure 7 depicts the signals resulting from applying u ( k ) to the system with the quantizers designed for M = 2 and n L = 2 . This figure shows that the output signals obtained by quantization y q ( k ) follow the ideal output signal y ( k ) pretty well and that the error between them is small in both cases. Also, the inputs u q ( k ) of the static quantizers are shown for comparison.
To further validate this observation, in Figure 8 there are shown the output signals of the system where the neural network quantizers were designed for M = 2 and n L = 4 , and in Figure 9 the ones for M = 8 , n L = 2 and n L = 4 . From these, we see that the proposed quantizer works well.
In addition, the result with the static quantizer q case and the result with the optimal dynamic quantizer case proposed in [5] are shown in Figure 10 for comparison. The value of the performance index for the static quantizer is E ( q ) = 7.9961 and that for the optimal dynamic quantizer is E ( Q N N R ) = 3.6002 . On the other hand, the performance of the proposed quantizer Q N N R for M = 2 and n L = 2 is E ( Q N N R ) = 3.73724 and that for M = 2 and n L = 4 is E ( Q N N R ) = 3.66764 . From this comparison result, we see that the proposed quantizer achieves higher performance than the static quantizer. Then, we find that the proposed quantizer is similar to the optimal dynamic quantizer, although the proposed quantizer is designed with the time series data of plant inputs and outputs, i.e., without the model information of the plant. Therefore, we can confirm that the neural network in the proposed quantizer captures the dynamics of the plant appropriately based on the time series data of the plant input and output.
The minimum values of the performance indexes in Equation (9), found by DE, are listed in Table 2. In addition, this table lists the average performance indexes and their standard deviation. The values in this table are divided according to their M, initialization method, n L and type (regression or classification). There are two initialization methods implemented: uniform random (Urand) and Xavier.
Drawing conclusions from this table by simple observation is difficult. For example, looking at the minimum values of E ( Q N N ) in the case of M = 2 , it is possible to say that Q N N C have better performance (smaller E ( Q N N ) ) than Q N N R in most cases. The average values not always corroborate this observation. For M = 8 , Q N N R has the smallest value of E ( Q N N ) in each case. However, there is no evidence that there is a significant difference in the performance of these types of quantizers. Therefore, the analysis of variance (ANOVA) is used to check if there are significant differences or not among these values.
Because many factors influence E ( Q N N ) , the 3-way ANOVA (ANOVA with three factors) is used. The considered factors are Type, initialization method (Init.) and number of layers n L . The categories of each factor are known as elements. For instance, the elements of the factor Type are R ( Q N N R : regression) and C ( Q N N C : classification). The M is not taken as a factor, because M = 8 gives smaller E ( Q N N ) s than M = 2 . Then, it is not necessary to check which one gives better results. The considered significance level is α = 0.05 . The goal is to determine if there is some statistical difference among the E ( Q N N ) ’s means of the design methods.
A summary of this test is shown in Table 3. The ANOVA test shows if there are significant differences among sets of data. When doing 3-way ANOVA, it is possible to see not only if there is a significant difference among elements of a factor but also combinations of elements of different factors. In this particular case, it will tell if there is a significant difference between Q N N R and Q N N C , and also it will tell if there are differences among the combinations of the quantizer types and the initialization methods. Then, the 3-way ANOVA test is run separately for M = 2 and M = 8 . For the case of M = 2 , the significant difference is found only for the initialization method. For the case of M = 8 the significant difference is found for all the factors and the combinations of them with the exception of the combination of the quantizer type and n L .
So far only one type of activation function h ( · ) , the sigmoid function, have been used in the hidden layers to build the neural networks. However, there are other activation functions that can be used. In this section two additional activation functions are considered: the hyperbolic tangent (tanh) and the Rectified Linear Unit ( ReLU ). These functions were defined by
tanh ( a ) = 1 exp ( 2 a ) 1 + exp ( 2 a ) ,
ReLU ( a ) = max ( a , 0 ) .
In addition, the limits l i in (17) for Xavier Uniform initializatin are given by
l i = 6 K i 1 + K i for h ( a ) = tanh ( a ) ,
l i = 6 K i for h ( a ) = ReLU ( a ) .
Several numerical simulations were performed to compare the performance of the neural network quantizers built with these functions. The settings of these simulations are the same as in the previous cases where h = sigm , but they were carried out only for M = 8 . These simulations were run N r u n = 50 times for each case. The results are summaries in Table 4.
As before, it is difficult to conclude from the table by simple observation. Therefore, the ANOVA test is used to analyze the data. In this case, four factors influence the results: h , initialization method, n L and quantizer type. However, because the influence of n L is understood the focus in this section will be in the factors: h , initialization method, quantizer type, and the interaction among each other. Therefore, the 3-way ANOVA general linear model of E ( Q N N ) versus quantizer type (Type), initialization method (Init) and activation function h is considered. The significance level used in this analysis is α = 0.05 . The analysis of variance showed that the statistical null hypothesis that all the means are the same was rejected for every single factor and the combination of them. This means that in each case there is at least one element that significantly differs from the others. The Tukey pairwise comparison is made to see the differences among the quantizer’s design elements.
The results of this test are summarized in Table 5. From the results, we see the following things. First, they tell that there is a significant difference between the Q N N R and Q N N C , and that Q N N R outperforms Q N N C . Second, they show that there is a difference between the initialization methods and that the Urand method exhibits better performance than the Xavier method. These results corroborate the ones shown in Table 3 previously obtained for M = 8 and h = sigm . Third, the table shows that the performances of the considered activation functions vary significantly, that the one with the best performance is h = tanh , and that the one with the lowest performance is h = ReLU .

5. Conclusions

This paper introduces the concept of neural network quantizers that are designed using a set of the inputs and outputs of the plant. These quantizers are aimed at systems in which the model of the plant is unknown or unreliable. They are constructed using feedforward neural networks and static quantizers. Two types of neural network quantizers are proposed: regression-based type Q N N R and classification based type Q N N C . In addition, a design method based on differential evolution is proposed for these quantizers.
By means of several numerical examples, it was found that both types of neural network quantizers are effective alongside with their DE based design method. Furthermore, many variations were considered in the construction of these quantizers. These variations are reflected in the number of quantization levels ( M = { 2 , 8 } ), in the number of layers of the network ( n L = { 2 , 4 } ), in the type of network initialization technique (Urand, Xavier), and in the hidden layers’ activation functions ( sigm , tanh, ReLU ). Several conclusions were reached based on the analysis of variance performed on the simulations results. Some of the most important is that the quantizers based on regression outperform the ones based on classification, that the best initialization method is the random uniform (Urand), and that the activation function that gives the best performance is tanh.

Author Contributions

Conceptualization, Y.M. and J.E.R.R.; methodology, Y.M. and J.E.R.R.; software, J.E.R.R.; validation, J.E.R.R.; formal analysis, J.E.R.R.; investigation, J.E.R.R.; data curation, J.E.R.R.; writing—original draft preparation, J.E.R.R.; writing—review and editing, J.E.R.R. and Y.M.; visualization, J.E.R.R. and Y.M.; project administration, Y.M.; funding acquisition, Y.M.

Funding

This research was partly supported by JSPS KAKENHI (16H06094) and a research grant from The Mazda Foundation.

Acknowledgments

The authors would like to thank Professor Kenji Sugimoto, Nara Institute of Science and Technology for his support.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Hespanha, J.P.; Naghshtabrizi, P.; Xu, Y. A Survey of Recent Results in Networked Control Systems. Proc. IEEE 2007, 95, 138–162. [Google Scholar] [CrossRef] [Green Version]
  2. Gupta, R.A.; Chow, M. Networked Control System: Overview and Research Trends. Ind. Electron. IEEE Trans. 2010, 57, 2527–2535. [Google Scholar] [CrossRef]
  3. Mahmoud, M.S.; Hamdan, M.M. Fundamental issues in networked control systems. IEEE/CAA J. Autom. Sin. 2018, 5, 902–922. [Google Scholar] [CrossRef]
  4. Azuma, S.; Sugie, T. Optimal dynamic quantizers for discrete-valued input control. Automatica 2008, 44, 396–406. [Google Scholar] [CrossRef]
  5. Azuma, S.; Sugie, T. Dynamic Quantization of Nonlinear Control Systems. IEEE Trans. Autom. Control 2012, 57, 875–888. [Google Scholar] [CrossRef] [Green Version]
  6. Okajima, H.; Sawada, K.; Matsunaga, N. Dynamic Quantizer Design Under Communication Rate Constraints. IEEE Trans. Autom. Control 2016, 61, 3190–3196. [Google Scholar] [CrossRef]
  7. Sawada, K.; Okajima, H.; Matsunaga, N.; Minami, Y. Dynamic quantizer design for MIMO systems based on communication rate constraint. In Proceedings of the 37th Annual Conference on IEEE Industrial Electronics Society, Melbourne, Austrilia, 7–10 November 2011; pp. 2572–2577. [Google Scholar] [CrossRef]
  8. Ramirez, J.E.; Minami, Y.; Sugimoto, K. Design of finite-level dynamic quantizers by using covariance matrix adaptation evolution strategy. Int. J. Innov. Comput. Inf. Control 2016, 795–808. [Google Scholar] [CrossRef]
  9. Ramirez, J.E.R.; Minami, Y.; Sugimoto, K. Synthesis of event-triggered dynamic quantizers for networked control systems. Expert Syst. Appl. 2018, 188–194. [Google Scholar] [CrossRef]
  10. Cybenko, G. Approximation by superpositions of a sigmoidal function. Math. Control Signals Syst. 1989, 303–314. [Google Scholar] [CrossRef]
  11. Hornik, K.; Stinchcombe, M.; White, H. Multilayer feedforward networks are universal approximators. Neural Netw. 1989, 359–366. [Google Scholar] [CrossRef]
  12. Moustakis, N.; Yuan, S.; Baldi, S. An adaptive approach to zooming-based control for uncertain systems with input quantization. In Proceedings of the 2018 European Control Conference (ECC), Limassol, Cyprus, 13–15 June 2018; pp. 2423–2428. [Google Scholar] [CrossRef]
  13. Liu, K.; Fridman, E.; Johansson, K.H. Dynamic quantization of uncertain linear networked control systems. Automatica 2015, 59, 248–255. [Google Scholar] [CrossRef]
  14. Ren, W.; Xiong, J. Quantized Feedback Stabilization of Nonlinear Systems With External Disturbance. IEEE Trans. Autom. Control 2018, 63, 3167–3172. [Google Scholar] [CrossRef]
  15. Takijiri, K.; Ishii, H. Networked control of uncertain systems via the coarsest quantization and lossy communication. Syst. Control Lett. 2018, 119, 57–63. [Google Scholar] [CrossRef]
  16. Almakhles, D.; Swain, A.; Nasiri, A.; Patel, N. An Adaptive Two-Level Quantizer for Networked Control Systems. IEEE Trans. Control Syst. Technol. 2017, 25, 1084–1091. [Google Scholar] [CrossRef]
  17. Rodriguez Ramirez, J.E.; Minami, Y.; Sugimoto, K. Neural network quantizers for discrete-valued input control. In Proceedings of the 2017 11th Asian Control Conference (ASCC), Gold Coast, Australia, 12–15 December 2017; pp. 2019–2024. [Google Scholar] [CrossRef]
  18. Rodriguez Ramirez, J.E.; Minami, Y.; Sugimoto, K. Design of Quantizers with Neural Networks: Classification Based Approach. In Proceedings of the 2018 International Symposium on Nonlinear Theory and Its Applications (NOLTA2018), Tarragona, Spain, 2–6 September 2018; pp. 312–315. [Google Scholar]
  19. Ojha, V.K.; Abraham, A.; Snášel, V. Metaheuristic design of feedforward neural networks: A review of two decades of research. Eng. Appl. Artif. Intell. 2017, 60, 97–116. [Google Scholar] [CrossRef] [Green Version]
  20. Ilonen, J.; Kamarainen, J.K.; Lampinen, J. Differential Evolution Training Algorithm for Feed-Forward Neural Networks. Neural Process. Lett. 2003, 93–105. [Google Scholar] [CrossRef]
  21. PÉrez, J.; Cabrera, J.A.; Castillo, J.J.; Velasco, J.M. Bio-inspired spiking neural network for nonlinear systems control. Neural Netw. 2018, 15–25. [Google Scholar] [CrossRef] [PubMed]
  22. Storn, R.; Price, K. Differential Evolution—A Simple and Efficient Heuristic for global Optimization over Continuous Spaces. J. Glob. Optim. 1997, 11, 341–359. [Google Scholar] [CrossRef]
  23. Price, K.; Storn, R.M.; Lampinen, J.A. Differential Evolution: A Practical Approach to Global Optimization, 1st ed.; Natural Computing Series; Springer: Berlin/Heidelberg, Germany, 2005. [Google Scholar]
  24. Kim, I.M.T.; Sugie, T. Fixed-structure controller synthesis: A meta-heuristic approach using simple constrained particle swarm optimization. Automatica 2009, 553–559. [Google Scholar] [CrossRef]
  25. Vapnik, V.N. Statistical Learning Theory; Wiley-Interscience: Hoboken, NJ, USA, 1998; Chapter 9; pp. 395–399. [Google Scholar]
  26. Alpaydin, E. Introduction to Machine Learning; The MIT Press: Cambridge, MA, USA, 2010; pp. 233–277. [Google Scholar]
  27. Glorot, X.; Bengio, Y. Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, Sardinia, Italy, 13–15 May 2010; Volume 9, pp. 249–256. [Google Scholar]
Figure 1. Considered system with a neural network quantizer Q N N .
Figure 1. Considered system with a neural network quantizer Q N N .
Electronics 08 00318 g001
Figure 2. Difference between the neural network quantizer based on regression and the one based on classification. (a) Regression based approach. The neural net has one output and shapes the original signal; (b) Classification based approach. The number of the neural network output is same as that of quantization levels, and each output correponds to the probability that a original signal is classified into a specific quantization level.
Figure 2. Difference between the neural network quantizer based on regression and the one based on classification. (a) Regression based approach. The neural net has one output and shapes the original signal; (b) Classification based approach. The number of the neural network output is same as that of quantization levels, and each output correponds to the probability that a original signal is classified into a specific quantization level.
Electronics 08 00318 g002
Figure 3. Example of a static quantizer q ( · ) for Q N N R ( M = 4 ).
Figure 3. Example of a static quantizer q ( · ) for Q N N R ( M = 4 ).
Electronics 08 00318 g003
Figure 4. Fully connected 3 layered feed-forward neural network example.
Figure 4. Fully connected 3 layered feed-forward neural network example.
Electronics 08 00318 g004
Figure 5. Example of static quantizer q [ · ] adapted for Q N N C ( M = 4 ).
Figure 5. Example of static quantizer q [ · ] adapted for Q N N C ( M = 4 ).
Electronics 08 00318 g005
Figure 6. Error system.
Figure 6. Error system.
Electronics 08 00318 g006
Figure 7. Signals resulting from applying u ( k ) to the system with the Q N N designed for M = 2 and n L = 2 . The black lines represent the signals without quantization ( u ( k ) , y ( k ) ) and the blues ones are the signals when quantization is applied ( v ( k ) , u q ( k ) , y q ( k ) ).
Figure 7. Signals resulting from applying u ( k ) to the system with the Q N N designed for M = 2 and n L = 2 . The black lines represent the signals without quantization ( u ( k ) , y ( k ) ) and the blues ones are the signals when quantization is applied ( v ( k ) , u q ( k ) , y q ( k ) ).
Electronics 08 00318 g007
Figure 8. Output signals y q ( k ) (blue) and y ( k ) (black) resulting from applying u ( k ) to the error system with Q N N designed for M = 2 and n L = 4 .
Figure 8. Output signals y q ( k ) (blue) and y ( k ) (black) resulting from applying u ( k ) to the error system with Q N N designed for M = 2 and n L = 4 .
Electronics 08 00318 g008
Figure 9. Output signals y q ( k ) (blue) and y ( k ) (black) resulting from applying u ( k ) to the error system with Q N N designed for M = 8 , n L = 2 (upper figure) and n L = 4 (lower figure).
Figure 9. Output signals y q ( k ) (blue) and y ( k ) (black) resulting from applying u ( k ) to the error system with Q N N designed for M = 8 , n L = 2 (upper figure) and n L = 4 (lower figure).
Electronics 08 00318 g009
Figure 10. Signals resulting from applying u ( k ) to the systems with the static quantizer q in Figure 3 and the optimal dynamic quantizer proposed in [5]. The black lines represent the signals without quantization ( u ( k ) , y ( k ) ) and the blues ones are the signals when quantization is applied ( v ( k ) , u q ( k ) , y q ( k ) ).
Figure 10. Signals resulting from applying u ( k ) to the systems with the static quantizer q in Figure 3 and the optimal dynamic quantizer proposed in [5]. The black lines represent the signals without quantization ( u ( k ) , y ( k ) ) and the blues ones are the signals when quantization is applied ( v ( k ) , u q ( k ) , y q ( k ) ).
Electronics 08 00318 g010
Table 1. Simulation conditions.
Table 1. Simulation conditions.
Quantizer TypeMK n L n w n
R: Q N N R { 2 , 8 } [ 10 , 1 ] 2132133
[ 10 , 10 , 10 , 1 ] 4352353
C: Q N N C 2 [ 10 , 2 ] 2142143
[ 10 , 10 , 10 , 2 ] 4362363
8 [ 10 , 8 ] 2208209
[ 10 , 10 , 10 , 8 ] 4428429
Table 2. E ( Q N N ) analysis for h = sigmoid ( N r u n = 50 ).
Table 2. E ( Q N N ) analysis for h = sigmoid ( N r u n = 50 ).
MInit. n L TypeMin.Avg.Std. Dev.
2Urand2R3.737244.329870.48300
C3.669464.270380.34762
4R3.667644.225160.45819
C3.547734.301580.49296
Xavier2R3.428794.158300.36054
C3.533074.096960.35038
4R3.650814.106350.35909
C3.638224.040660.29597
8Urand2R0.178250.206220.01612
C0.912012.911111.54333
4R0.229110.322430.14041
C0.816672.816301.43382
Xavier2R0.204241.900451.22825
C0.932842.855801.44467
4R0.247620.851881.03043
C1.000162.612011.06714
Table 3. Tukey pairwise comparison for h = sigmoid and single factors]Tukey pairwise comparison 3-way ANOVA for h = sigmoid and single factors. Grouping information using the Tukey test and 95 % confidence. Means that do not share a letter are significantly different.
Table 3. Tukey pairwise comparison for h = sigmoid and single factors]Tukey pairwise comparison 3-way ANOVA for h = sigmoid and single factors. Grouping information using the Tukey test and 95 % confidence. Means that do not share a letter are significantly different.
MFactorNMeanGrouping
2TypeR2004.20492A
C2004.17739A
InitUrand2004.28175A
Xavier2004.10057 B
n L L22004.21388A
L42004.16844A
8TypeC2002.79881A
R2000.82024 B
InitXavier2002.05504A
Urand2001.56401 B
n L L22001.96839A
L42001.65066 B
Table 4. E ( Q N N ) results summary for h = tanh and h = ReLU ( M = 8 ).
Table 4. E ( Q N N ) results summary for h = tanh and h = ReLU ( M = 8 ).
h Init. n L TypeMin.Avg.Std. Dev.
tanhUrand2R0.179450.573430.73814
C0.857072.170401.21598
4R0.243401.836211.37026
C0.696041.621320.68768
Xavier2R0.170100.207610.02103
C0.929431.970480.78436
4R0.198680.253250.03900
C0.660411.771380.99609
ReLU Urand2R0.165320.758541.21252
C0.704941.932260.98668
4R0.191753.041531.69148
C0.717181.938231.27409
Xavier2R0.224382.972021.24661
C0.723982.129151.12578
4R0.220623.016701.70763
C0.631882.143191.03974
Table 5. Tukey pairwise comparison 3-way ANOVA for the activation functions comparison ( M = 8 ). Grouping information using the Tukey test and 95 % confidence. Means that do not share a letter are significantly different.
Table 5. Tukey pairwise comparison 3-way ANOVA for the activation functions comparison ( M = 8 ). Grouping information using the Tukey test and 95 % confidence. Means that do not share a letter are significantly different.
FactorNMeanGrouping
TypeC6002.23930A
R6001.32836 B
InitXavier6001.89033A
Urand6001.67733 B
h ReLU4002.24145A
sigm4001.80953 B
tanh4001.30051 C

Share and Cite

MDPI and ACS Style

Rodriguez Ramirez, J.E.; Minami, Y. Design of Neural Network Quantizers for Networked Control Systems. Electronics 2019, 8, 318. https://doi.org/10.3390/electronics8030318

AMA Style

Rodriguez Ramirez JE, Minami Y. Design of Neural Network Quantizers for Networked Control Systems. Electronics. 2019; 8(3):318. https://doi.org/10.3390/electronics8030318

Chicago/Turabian Style

Rodriguez Ramirez, Juan Esteban, and Yuki Minami. 2019. "Design of Neural Network Quantizers for Networked Control Systems" Electronics 8, no. 3: 318. https://doi.org/10.3390/electronics8030318

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop