Next Article in Journal
The Relative Performance of a Benchtop Scanning Monochromator and Handheld Fourier Transform Near-Infrared Reflectance Spectrometer in Predicting Forage Nutritive Value
Previous Article in Journal
A Method for Cheating Indication in Unproctored On-Line Exams
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Flatness Prediction of Cold Rolled Strip Based on Deep Neural Network with Improved Activation Function

1
College of Sciences, Northeastern University, Shenyang 110819, China
2
College of Information Science and Engineering, Northeastern University, Shenyang 110819, China
*
Authors to whom correspondence should be addressed.
Sensors 2022, 22(2), 656; https://doi.org/10.3390/s22020656
Submission received: 21 November 2021 / Revised: 23 December 2021 / Accepted: 11 January 2022 / Published: 15 January 2022
(This article belongs to the Section Physical Sensors)

Abstract

:
With the improvement of industrial requirements for the quality of cold rolled strips, flatness has become one of the most important indicators for measuring the quality of cold rolled strips. In this paper, the strip production data of a 1250 mm tandem cold mill in a steel plant is modeled by an improved deep neural network (the improved DNN) to improve the accuracy of strip shape prediction. Firstly, the type of activation function is analyzed, and the monotonicity of the activation function is deemed independent of the convexity of the loss function in the deep network. Regardless of whether the activation function is monotonic, the loss function is not strictly convex. Secondly, the non-convex optimization of the loss functionextended from the deep linear network to the deep nonlinear network, is discussed, and the critical point of the deep nonlinear network is identified as the global minimum point. Finally, an improved Swish activation function based on batch normalization is proposed, and its performance is evaluated on the MNIST dataset. The experimental results show that the loss of an improved Swish function is lower than that of other activation functions. The prediction accuracy of a deep neural network (DNN) with an improved Swish function is 0.38% more than that of a deep neural network (DNN) with a regular Swish function. For the DNN with the improved Swish function, the mean square error of the prediction for the flatness of cold rolled strip is reduced to 65% of the regular DNN. The accuracy of the improved DNN is up to and higher than the industrial requirements. The shape prediction of the improved DNN will assist and guide the industrial production process, reducing the scrap yield and industrial cost.

1. Introduction

Cold rolled strips with high-quality surface, dimension, flatness, and mechanical properties are used in construction, automobiles, household appliances, packaging, building materials, etc., the output and quality level of which reflects the technological strength of a country’s iron and steel industry [1]. With the improvement of industry requirements for the quality of cold rolled strips, the quality of strip shape has become one of the most important indexes to evaluate cold rolled strips. The strip shape refers to the warpage degree of the plate, and mass refers to the distribution of residual stress in the strip. The composition of the strip shape includes the cross-sectional geometry and flatness of the strip in a natural state. Therefore, the quantitative description of plate shape involves multiple indicators in these two aspects, including convexity, wedge, edge thinness, local high points, and flatness [2].
The setting and calculation of strip shape are related to the rolling force and roll bending force required for the rolling deformation of the steel strip. The change of the bending force for the work roll and intermediate roll could deform the roll and change the value of the roll gap; the change of roll gap directly affects the flatness of the strip. Tilt and translation could cause pressure to change the roll, which also affects flatness. Claire et al. [3] proposed an effect function to describe the adjustment effect of the unit adjustment for the adjustment method of the flatness deviation. Wang et al. [4] enriched the effect function in the control system for the flatness of the cold rolling mill and proved that the defined effect function and distribution of flatness deviation were consistent in the form of expression. Chen et al. [5] established the finite element model of a six-high rolling mill based on the effect function to calculate the horizontal and vertical stiffness of the rolling mill. They obtained the influence of the change in values of the no-load roll gap and the bending force of the work roll and intermediate roll on the exit thickness and exit plate crown, which improved the lack of a priori coefficient. However, the effect function could not reflect the flatness in real-time. Therefore, the real-time control model of flatness was proposed by calculating the real-time strip shape during the production process based on the control coefficient of the model. Liang et al. [6] proposed a flatness feed-forward control model which adjusted the corresponding compensation value for the roll bending force by changing the rolling force, regarding it as the pre-control of the closed-loop feedback control system for flatness. The sum of the squares, regarding the difference between the change of the rolling force and the adjustment of the pressure distribution in the loading gap between the work roll bending roll and the intermediate roll bending roll pair, was minimized by minimizing the objective function, thereby controlling the flatness. To make the objective function smaller, Wang et al. [7] proposed a shape control efficiency coefficient based on the feed-forward control model, which refers to a unit change in the shape of the bearing roll gap caused by the effect of the shape actuator. This can be used to calculate the elastic deformation value of the roller system via the influence function method and the difference method. Liu et al. [8] proposed a fuzzy neural network based on a dynamic efficiency matrix to predict control efficiency coefficients. Based on obtaining the control efficiency coefficient by putting the control means into the neural network. This network was proposed to use the rolling force, strip width, actual strip flatness deviation, and actual position change of the actuator as the input, and the efficiency coefficient of 20 control points as the output to train the shallow neural network, further improving the prediction accuracy [9]. However, deep learning (DL) brought new research methods to the study of strip shape. Surface defect detection [10] was also an important research direction, involving a convolutional neural network (CNN) [11,12,13,14,15,16], deep belief network (DBN) [17], visual geometry group network (VGG) [18], and the combination of DL and an extreme learning machine [19,20,21].
The control of strip shape is a complex physical process, and there is a large amount of data in the strip production process. Data in the strip tandem cold rolling process is combined with deep learning technology to establish a predictive model. Through the self-learning of data, many hidden and complex knowledge patterns were obtained, which benefited from investigations regarding the influence of numerical change on the flatness control means of the work roll and intermediate roll, such as the bending force, roll tilt, and roll traverse, on the final exit shape during the tandem cold rolling process. The transformed adjustment of the strip shape during the production process facilitates pre-production control and regulation regarding the input prediction of the strip shape. By changing the input value continuously, more ideal results of flatness prediction are obtained, and the exit shape is controlled effectively. The model could give full weight to the value of data accumulated on the production line, realize data-driven production, and forecast predictions before the start of production. According to the target flatness, each value of the regulation mean is adjusted in advance, reducing the adjustment required during the production process. Therefore, production costs could be saved, and the quality of the flatness and the yield rate of the strip could be improved. In addition, the monotonicity of the activation function is proven unnecessary through the non-convex loss function in the deep learning algorithm. In this paper, the Swish activation function is further improved and optimized, the training and calculation speed of the model is increased, and a higher precision of prediction is realized. Python was used to obtain all of the results and graphs in this paper.
The rest of this paper is organized as follows: Section 2 presents a brief review of the basic concepts and related work of the deep neural network, Section 3 describes the proposed improved deep neural network, Section 4 reports and analyzes the experimental results, and Section 5 summarizes the conclusion.

2. Related Theory

2.1. Deep Neural Network

A deep neural network(DNN) is an architectural model in deep learning. A feed-forward neural network is composed of an input layer, multiple hidden layers, and an output layer. The layers of the DNN and the neurons between the layers are fully connected; however, the neurons within the layers are not connected. The training of the DNN is divided into forwarding propagation calculation and backpropagation derivation.
Suppose the input samples are X = ( X 1 , X 2 , , X n ) , each sample has m features, and the actual output Y = ( y 1 , y 2 , , y n ) is known. Suppose the output of the l-th layer is a l and the input layer is the 0-th layer. This could obtain a 0 = X . Each hidden layer contains two parameters, namely the weight W and the bias b. The weight W is a matrix whose dimension is determined by the output dimension j of the previous layer and the number of hidden units k in this layer; that is, W is a matrix with j rows and k columns. The b is a 1 × k vector. The output layer, the same as the hidden layer, contains these two parameters. Each layer will calculate the output of the previous layer. The calculation is divided into two parts. The calculation of the l-th layer is:
Z l = ( W l ) T a l 1 + b l
a l = σ ( Z l )
where Wl is the weight of the l-th layer, T represents the transpose of a matrix, bl is the bias of the l-th layer, al−1 is the output of the previous layer, and σ is the activation function. The output al calculated by the l-th layer is regarded as the input of the next layer for calculation until the output layer calculates the predicted value y ^ . The calculation process of the output layer is the same as that of the hidden layers.
The forward propagation calculation of a sample is completed, and the next step is the backward propagation derivation: the forward propagation calculation obtains the predicted value y ^ , which has some difference with the sample’s true y value; the difference is measured by the loss function L which needs to be selected according to the data type and the final purpose, then it can be expressed as L ( y , y ^ ) . The goal of DNN training is to minimize the difference and improve the prediction accuracy of the final model. Therefore, the goal of backpropagation is to update the parameters of the previous layers to continuously reduce the error.

2.2. Type of Activation Function

The activation function is an important part of the neural network. The activation function of a hidden layer is generally the same from layer to layer. The main purpose of the activation function is to provide nonlinear modeling ability.
a l = ( W l ) T a l 1 + b l
a l + 1 = ( W l + 1 ) T a l + b l + 1   = ( W l + 1 ) T ( ( W l ) T a l 1 + b l ) + b l + 1   = W ˜ a l 1 + b ˜
No matter how many layers of the neural network are calculated, if there is no activation function, the final output result is still the same, W ˜ a 0 + b ˜ . The multi-layer calculation of neural networks is equivalent to that of only one layer. Therefore, DNN has the ability of hierarchical nonlinear mapping with the addition of the activation function [19,20]. The common activation functions are shown in Table 1.

2.3. Loss Function

The loss function is used to measure the difference between the predicted value y ^ and the true value y. On the one hand, the loss function is used in the training process of the model. The loss function is also used as the objective function of the model and through the optimization method, continuously reduces the loss function to update the network parameters to achieve the optimization process of the objective function. On the other hand, it is used to evaluate the prediction results of the final model after training. The smaller the loss value, the better the prediction effect and the higher the accuracy.
During the regression of the DNN, the mean square error loss function is generally used as the objective function to train the model, and MSE is used to evaluate the final model after training. In the classification model of DNN, the cross-entropy loss function is generally used as the objective function to train the model and precision and recall are used to evaluate the classification results. The mean square error (MSE) loss function is used in this study, and its implementation formula is as follows:
L ( y , y ^ ) = 1 n i = 1 n ( y i y ^ i ) 2

3. Improved Deep Neural Network

The activation function is an indispensable core part of the neural network. Regardless of whether the neural network has a DNN, RNN, or convolutional neural network (CNN) structure, the hidden layer and the output layer require an activation function, making the neural network capable of nonlinear modeling. First, it is concluded that the monotonicity of the activation function in the DNN is unnecessary by analyzing the non-convexity of the loss function (Appendix A). When the non-convex loss function can take the minimum value, it will be discussed (Appendix B). Then the advantages of the ELU and Swish function were integrated to improve the Swish function. A performance test was facilitated to investigate the improved activation function on a scientific dataset.
The motivation for improving the activation function comes from the idea of batch normalization. Batch normalization standardizes the output of each layer of the DNN; the mean value of the output is 0, and the variance is 1.
u = 1 n i = 1 n z ( i )
σ 2 = 1 n i = 1 n ( z ( i ) u ) 2
z n o r m a l i z a t i o n ( i ) = z ( i ) u σ 2 + ε
The advantages of batch normalization are as follows: First, it reduces the deviation of the internal covariates for the output of each layer in the network, making the update of weight more robust, especially in deep neural networks. Therefore, the weight of the back layer for the network is more inclusive of the weight of the front layer. This means that the weight change of the front layer has little effect on the weight of the back layer, and the overall network is more robust; thereby the generalization ability of the network can be improved. Secondly, it normalizes the output of each layer so that the gradient is in an unsaturated region, which can effectively avoid the problems of gradient explosion and vanishing gradient. Finally, the consistency of the overall weight update is maintained, which effectively accelerates the network training process.
Combining the advantages of the ELU activation function, the non-monotonic activation function, Swish, is improved as follows.
f ( x ) = { x , x 0 λ x e x , x < 0
The improved Swish activation function is non-monotonic. On the positive half-axis of the x-axis, the gradient remains at 1 and greater than the gradient of the Swish function, making the gradient drop faster. On the negative half-axis of the x-axis, the absolute value of the negative gradient is increased so that a faster gradient descent speed is maintained on the negative semi-axis, and the output on the negative semi-axis is negative. The parameter λ is used to adjust the magnitude of the negative output, and the average output value is 0, which speeds up training.
A neural network with an 0 output mean value activation function can be called a self-normalized neural network, which realizes the batch normalization through the activation function, making the training process easier. There is no need to extract the output of each layer for normalization to reduce the error rate of the operation and simplify the calculation. In the improved Swish activation function, the method of batch normalization is used for reference, and the specific value of the parameter λ is calculated. By making the mean value for the output of the improved Swish activation function 0, the effect of batch normalization is achieved. Assuming that the input of each layer is a standard normal distribution with a mean value of 0 and a variance of 1, the initial value for the weight of each layer obeys the standard normal distribution, and a normal random number is generated to obtain λ = 2.9046 . Therefore, the improved Swish activation function formula is as follows.
f ( x ) = { x ,       x 0 2.9046 x e x , x < 0
The improved Swish activation function is derivable when x 0 . The derivative of the function is set to 1 when x = 0 . The improved Swish activation function draws on the advantages of the ELU and Swish functions and satisfies the conditions as the activation function of the DNN while maintaining a large gradient. It can speed up the training and make the mean value of output 0 to achieve batch normalization. The effect speeds up the training speed while preventing gradient explosion and disappearance, exerting a better effect in the DNN.

4. Experimental Results and Discussion

4.1. Description of Data

The dataset is the measured data in the actual production of the cold rolling mill, which comes from the five stands in the first production sequence. The established model aimed to predict the final export shape of cold rolled strip steel through various influencing factors. Therefore, the work roll bending force, intermediate roll bending force, rolling force, tension, crimp tension, intermediate roll transverse displacement of the five stands in the first production sequence, and the export flatness measured in each sensor area of the first stand are taken as the input variables, and the export flatness measured in each sensor area of the fifth stand is taken as the output variables.
After deleting the invalid data, empty data, and scrambled data in the dataset, there are 234,527 rows and 86 columns of valid data, namely 234,527 sample points. These sample points are obtained according to time measurement, and a set of data is obtained every 0.08 s. The measured dataset may also continue to change when all known variables are almost unchanged because these influencing factors have a continuous impact on the flatness of the plate. Therefore, time is also an important factor. A column of time should be added to the dataset to reflect the change of sample points over time. Since the first 50,000 data were generated in the unstable rolling process and did not belong to the prediction category of sub-modeling, the first 50,000 data were excluded.
After data cleaning and processing, the dataset is firstly segmented. A total of 184,527 data samples are randomly assigned to the training set and the test set according to the ratio of 7:3. Then, the input data is standardized.
Due to the uniform distribution of data in each column, the corresponding w i obtained by training has a small numerical difference. The gradient descent surface of the loss function along the change of w and b resembles a round bowl. When the loss function is gradient descent, the objective function can also ensure continuous decline, and no oscillation occurs if the learning rate α is large. The gradient descent step becomes larger, and the loss function decreases rapidly, then the training speed can be accelerated.

4.2. Experiments Based on MNIST Dataset

The MNIST dataset is a handwritten digitized database with 60,000 examples in the training library and 10,000 examples in the test library. It is a standard scientific dataset for evaluating the performance of an algorithm by the accuracy of the training set and test set and the decline of the loss function. The MNIST dataset is used to evaluate the performance of the improved Swish activation function and compare it with Sigmoid, ReLU, Swish, and ELU activation functions.
Figure 1 and Figure 2 compare the loss and accuracy of deep learning models with the improved Swish activation function, Sigmoid function, and ReLU function respectively. The improved Swish and ReLU functions show similar performance in loss reduction and accuracy improvement. However, the improved Swish function’s loss reduction and accuracy improvement are completely superior to the Sigmoid function both in training and test steps.
Figure 3 and Figure 4 compare the loss and accuracy of the deep learning model with the improved Swish, ELU, and original Swish functions, respectively. In terms of model loss, the model with the improved Swish activation function was similar to the training set and test set, compared with the model of the ELU and Swish function; For accuracy, the accuracy of the model with the improved Swish activation function in the training set is slightly lower than that of the model with original Swish function, but higher than that of the model with ELU function, while the accuracy of the model with the improved Swish activation function in the test set is higher than that of the model with the original Swish function and ELU function.
The loss and accuracy of the models with different activation functions on the training set and the test set are shown in Table 2.
The accuracy of the model with the improved Swish activation function on the test set is higher than that of the models with other activation functions. The accuracy on the training set is similar to the original Swish and ReLU function but higher than the ELU and Sigmoid function. The loss of the model using the improved Swish activation function on the test set is smaller than other activation functions. Compared with the current common activation functions, the improved Swish activation function model has achieved good results on the MNIST dataset, and it has better performance in a deep neural network.

4.3. Flatness Prediction of Cold Rolled Strip

The DNN with the improved Swish activation function is used to model the cold rolled strip data and analyze the prediction results. The datasets used in the modeling process and the data preprocessing process are detailed in Section 4.1. A total of 20 models are trained, and the model with the smallest MSE in the training set is selected as the optimal model (Figure 5).
In the DNN model with the improved Swish activation function, the MSE loss value of the training set and the test set are 1.281 and 1.305, respectively. The error between them is only 0.024; therefore, the model did not over-fit. Figure 5 show the decline of loss of the model during 1200 iterations. The loss function decreases rapidly in the first 50 iterations, and there is a slight oscillation between 50 and 100 times. After 100 times iteration, the loss value decreased smoothly and then finally converged.
The DNN with the improved Swish activation function was used to predict the flatness of the cold rolled strip, and the predicted value y ^ and true value y in each sensor area are compared in Table 3. The difference between the predicted value and the true value of the flatness in the f939 sensor area is large, and the difference in other sensor areas is small. An error of less than 5 % error is the smallest in the f947 sensor area, which is 0.008 %. The MSE of the difference in each sensor area is 1.166, and the prediction accuracy is high.
Figure 6 show the comparison between the predicted and true values of the data of the f9314 sensor area. The numerical fitting effect of the DNN with the improved Swish activation function (Improved DNN) in 36 sensor regions is shown in Figure 7. The DNN with the improved Swish activation function has a good fitting ability for the true value, and its predicted value reflects the true value.
The BP, DNN, and DNN with the improved Swish activation function were used to predict the flatness of cold rolled strips (Table 4). MSE of the above models on the training set and test set shows a downward trend, and the DNN model and the DNN with the improved Swish activation function are superior to the BP in terms of prediction ability. Compared with DNN, the DNN with the improved Swish activation function has smaller MSE and higher prediction accuracy. The model can predict the flatness of cold rolled strips accurately and higher than the industrial requirements. The flatness prediction results can assist and guide the industrial production process, reduce the yield of scrap, and reduce the industrial cost.

5. Conclusions

In the paper, it is concluded that the non-monotone function can be selected as the activation function by proving the non-convexity of the loss function in the DNN model and the proof that the critical point in the nonlinear deep neural network is the global minimum point was given. Then the parameters of the improved Swish activation function were calculated using the fixed point method and combining the advantages of the ELU and Swish activation functions. The experiments showed that the prediction accuracy of the MNIST dataset is improved by 0.38%. Finally, the DNN with the improved Swish activation function was used to model the actual production data of cold rolled strips. The mean square error for the flatness of cold rolled strip based on the deep neural network with the improved Swish activation function was reduced to 65% of that of the deep neural network. Using this model to predict the export shape of the cold rolled strip can be used to guide the actual production process of cold rolled strip, which is conducive to reducing the scrap rate of the strip and thus reducing the production cost.

Author Contributions

Conceptualization, J.L. and S.S.; methodology, J.L. and J.W.; software, J.W.; validation, J.W.; formal analysis, J.L. and J.W.; investigation, J.L.; resources, S.S.; data curation, J.W.; writing—original draft preparation, J.L. and J.W.; writing—review and editing, J.L., M.B., N.S. and S.L.; visualization, J.W., M.B., N.S. and S.L.; supervision, S.S.; project administration, S.S.; funding acquisition, J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Key Research and Development Program of China under Grant 2017YFB0304100, in part by the National Natural Science Foundation of China under Grant 71672032, in part by the Fundamental Research Funds for Central University under Grant N2005011, and in part by the Scientific Research Funds of Educational Department of Liaoning Province under Grant LT2020008.

Data Availability Statement

The authors thank the anonymous reviewers for their valuable comments to improve the paper quality.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Non-Convexity of the Loss Function

The significance of the monotonic activation function is that a single-layer network can make the loss function with a monotonic activation function a convex function [22]. In the optimization of the convex function, the local minimum would be the global minimum. When using an optimization algorithm such as gradient descent, the local-optimal solution obtained by convergence is the global-optimal solution. After multiple activation functions in the DNN, the loss function could not be strictly convex no matter whether the activation function is monotonous or not. The specific proof is given below:
First, the paper considers a special two-layer DNN. Suppose the activation function is f ( z ) = z , the weights are w 1 , w 2 , w y , then the output is:
f ( w 1 , w 2 , w y ) = x w 1 w 2 w y
Suppose for the loss function l o s s ( f ( w 1 , w 2 , w y ) ) , the paper obtains the optimal weight ( w a , w b , w y ) . The strictly convex function has only one optimum point. If the loss function l o s s ( f ( w 1 , w 2 , w y ) ) is a convex function, for any other weight ( w 1 , w 2 , w y ) ,
l o s s ( f ( w 1 , w 2 , w y ) ) > l o s s ( f ( w a , w b , w y ) )
Because of
f ( w 1 , w 2 , w y ) = x w a w b w y = x w a I w b w y = x w a A A 1 w b w y = x w 1 w 2 w y
When A and A 1 are elementary transformations, there are a series of weights ( w 1 , w 2 , w y ) to make l o s s ( f ( w 1 , w 2 , w y ) ) take the minimum value. To show whether ( w 1 , w 2 , w y ) and ( w a , w b , w y ) are the same point, the following lemma is given.
Lemma A1.
For matrix A, A is a zero matrix if there is any full-rank matrix B satisfies AB = A.
Proof of Lemma A1.
When A is a 1 × 1 matrix, that is A , B are constants, B 0 . If B is a full-rank matrix, then A = O .
It can be assumed that Lemma 1 holds when A is a k × k matrix. When A is a ( k + 1 ) × ( k + 1 ) matrix, A B = A can be expressed as a block matrix.
( A k k A k 1 A 1 k A 11 ) ( B k k B k 1 B 1 k B 11 ) = ( A k k B k k + A k 1 B 1 k A k k B k 1 + A k 1 B 11 A 1 k B k k + A 11 B 1 k A 1 k B k 1 + A 11 B 11 ) = ( A k k A k 1 A 1 k A 11 )
where A = O is one solution.
For any full-rank matrix B , B k 1 = O , B 1 k = O , so B 11 O . Formula (A1) can be written as:
( A k k B k k A k 1 B 11 A 1 k B k k A 11 B 11 ) = ( A k k A k 1 A 1 k A 11 )
when A k k = O and B 11 O , A k 1 = O , A 11 = O .
Considering A 1 k B k k = A 1 k , the expression can be:
( A 1 k O ( k 1 ) k ) B k k = ( A 1 k O ( k 1 ) k )
From the above assumptions, Lemma A1 holds when A is a k × k matrix, then A 1 k = O .
When A is a ( k + 1 ) × ( k + 1 ) matrix, Lemma A1 holds as well. By mathematical induction, Lemma 1 is proved.
According to Lemma A1, one of the following two situations must be satisfied:
(1) There is a full-rank matrix A such as w a w a A , that is w a w 1 . Then ( w a , w b , w y ) is not the only best point, and the loss function is not strictly convex.
(2) Any full-rank matrix A makes w a = w a A or w b = w b A , that is w a or w b a zero matrix. No matter what value the remaining parameter w y takes, the output of the network is 0. If it is the optimal value of ( w 1 , w 2 , w y ) , they are all optimal values for any w y . And ( w 1 , w 2 , w y ) is not the only optimal value, and the loss function is not strictly convex.
Then the paper considers the ordinary multi-layer neural network. Assuming that the activation function is an arbitrary function α ( z ) and A is an elementary transformation,
f ( w 1 , w 2 , w y ) = α ( x w a ) w b w y = α ( x w a ) I w b w y = α ( x w a ) A A 1 w b w y = α ( x w a A ) A 1 w b w y = α ( x w 1 ) w 2 w y
In the same way, the loss function is not strictly convex.
In summary, the loss function of the multi-layer neural network is not strictly convex regardless of the monotonicity of the activation function. Since the loss function of DNN is non-convex, the activation function does not require monotonicity, and a non-monotonic function can be selected as the activation function. □

Appendix B

Global Minimum of the Loss Function

It is difficult to solve the optimal solution for non-convex optimization problems. Although the loss function of DNN is not convex, the critical point is more likely to be a saddle point rather than a false local minimum. Most of the local minimums of a multilayer neural network are at the bottom of the surface for the loss image; it is close to the global minimum. No matter whether it is a local minimum or a global minimum, the result is acceptable when the loss function falls within the range the paper requires. It can be proven that all local minima of the nonlinear deep neural network are global minima under the condition of full rank.
Suppose there are a total of m sample points, the input vector is X R d x , output vector is Y R d y , the network has H hidden layers, the width of each layer is d 1 , , d H , and d 0 = d x , d H + 1 = d y . Let the weights from the k−1 layer to the k layer be W k R d k × d k 1 , k = 1 , , H + 1 . Let σ min ( A ) and σ max ( A ) be the minimum and maximum singular values of matrix A respectively, and the activation function is ψ ( x ) .
L = 1 m i = 1 m ψ ( W H + 1 ψ ( W 2 ψ ( W 1 X ( i ) ) ) )
When a smooth nonlinear function h is given, h maps the multi-dimensional input to the multi-dimensional output. If each layer of the network is set as a smooth nonlinear function h i : R d i 1 R d i , i = 1 , , H + 1 , the entire network can be expressed as a combination of h i , that is, h can be decomposed into a series of smooth nonlinear functions, namely
h ( H + 1 ) : 1 = h H + 1 h H h 1
Therefore, the goal of deep neural network learning is to find a set of nonlinear functions h 1 , h 2 , , h H + 1 to minimize the overall loss function.
L ( h ) = L ( h 1 , h 2 , , h H + 1 ) = 1 2 E [ h H + 1 h H h 1 ( X ) Y 2 2 ]
The minimization of the square loss is the conditional expectation of Y given X, that is h ( x ) = E [ Y | X = x ] , which divides the loss function into two parts.
L ( h ) = 1 2 E [ h H + 1 h 1 ( X ) h ( X ) 2 2 ] + C
where C is a constant term, representing the deviation independent of h 1 , h 2 , , h H + 1 . If h H + 1 h H h 1 = h , the optimal solution L = C is obtained at this time.
The following can prove that Theorem 1 holds when the following conditions are met.
(1) Suppose the function space satisfies F = { h : R d x R d y | h ( 0 ) = 0 , sup x h ( x ) 2 x 2 < } , where h F can be derived, F i = { h i : R d i 1 R d i | h i ( 0 ) = 0 , sup x h i ( x ) 2 x 2 < } , i = 1 , , H + 1 , where h 1 F 1 , , h H + 1 F H + 1 can be derived and h H + 1 h 1 F .
(2) Suppose d x d y , d i d x , d i d y , i = 1 , 2 , , H + 1 .
Theorem A1.
If  h H + 1 : 2 ( Z ) is secondarily differential and  ε > 0 such that  σ min ( J [ h H + 1 : 2 ] ( Z ) ) ε holds for any  Z R d 1 , the critical point of  L ( h ) is also the global minimum.
Proof of Theorem A1.
For η F i , by the definition of Frechet derivative,
D h i [ L ( h ) ] , η = lim ε 0 L ( h 1 , , h i + ε η , , h H + 1 ) L ( h ) ε
where D h i [ L ( h ) ] represents the Frechet derivative of the overall loss h i , and D h i [ L ( h ) ] , η represents the directional derivative along the direction η . Let J [ f ] ( x ) denote the Jacobian matrix of function f at point x, then
L ( h 1 , , h i + ε η , , h H + 1 ) = 1 2 E [ h H + 1 : i + 1 ( h i + ε η ) h i 1 : 1 ( X ) h ( X ) 2 2 ] + C = 1 2 E [ h H + 1 : i + 1 ( h i : 1 ( X ) + ε η ( h i 1 : 1 ( X ) ) ) h ( X ) 2 2 ] + C = 1 2 E [ h H + 1 : 1 ( X ) + ε J [ h H + 1 : i + 1 ] ( h i : 1 ( X ) ) η ( h i 1 : 1 ( X ) ) + O ( ε 2 ) h ( X ) 2 2 ] + C = L ( h ) + ε E [ ( h H + 1 : 1 ( X ) h ( X ) ) T J [ h H + 1 : i + 1 ] ( h i : 1 ( X ) ) η ( h i 1 : 1 ( X ) ) ] + O ( ε 2 )
Then
D h i [ L ( h ) ] , η = E [ ( h H + 1 : 1 ( X ) h ( X ) ) T J [ h H + 1 : i + 1 ] ( h i : 1 ( X ) ) η ( h i 1 : 1 ( X ) ) ]
In the Formula (A1), the paper takes i = 1 to obtain
D h 1 [ L ( h ) ] , η = E [ ( h H + 1 : 1 ( X ) h ( X ) ) T J [ h H + 1 : 2 ] ( h 1 ( X ) ) η ( X ) ] η F 1
Suppose A ( X ) = J [ h H + 1 : 2 ] ( h 1 ( X ) ) , and A ( X ) is a matrix with row full rank from the hypothesis (2). Then A ( X ) A ( X ) T is invertible.
Define a special direction
η ˜ ( X ) = A ( X ) T ( A ( X ) A ( X ) T ) 1 ( h H + 1 : 1 ( X ) h ( X ) )
Then
D h 1 [ L ( h ) ] , η ˜ = E [ h H + 1 : 1 ( X ) h ( X ) 2 2 ]
Prove η ˜ ( X ) F 1 below.
If h H + 1 : 1 ( 0 ) h ( 0 ) = 0 , η ˜ ( 0 ) = 0 . If J [ h H + 1 : 2 ] is derivable and h 1 F 1 , A ( X ) , A ( X ) T , ( A ( X ) A ( X ) T ) 1 is derivable. If h H + 1 : 1 h F , η ˜ ( X ) is derivable.
Suppose Singular Value Decomposition (SVD) of A ( X ) is decomposed into A ( X ) = U V T , where = [ 1 O ] . Then
A ( X ) T ( A ( X ) A ( X ) T ) 1 = V T U T ( U V T V T U T ) 1 = V T U T ( U 1 2 U T ) 1 = V T U T U 1 2 U T = V [ 1 1 O ] U T
Suppose the operator norm of matrix A is A O P = sup x A x 2 x 2 , then
A ( X ) T ( A ( X ) A ( X ) T ) 1 O P = σ max ( A ( X ) T ( A ( X ) A ( X ) T ) 1 ) 1 ε
For h F , the paper defines the norm of the nonlinear function h as h n l = sup x h ( x ) 2 x 2 ,
η ˜ ( X ) 2 = A ( X ) T ( A ( X ) A ( X ) T ) 1 ( h H + 1 : 1 ( X ) h ( X ) ) 2 A ( X ) T ( A ( X ) A ( X ) T ) 1 O P h H + 1 : 1 ( X ) h ( X ) 2 A ( X ) T ( A ( X ) A ( X ) T ) 1 O P h H + 1 : 1 h n l X 2
Then
η ˜ n l A ( X ) T ( A ( X ) A ( X ) T ) 1 O P h H + 1 : 1 h n l h H + 1 : 1 h n l ε
Then η ˜ F 1 . Because of:
D h 1 [ L ( h ) ] O P D h 1 [ L ( h ) ] , η ˜ η ˜ n l ε E [ h H + 1 : 1 ( X ) h ( X ) 2 2 ] h H + 1 : 1 h n l = ε ( | L ( h ) L | ) h H + 1 : 1 h n l
then
D h 1 [ L ( h ) ] O P h H + 1 : 1 h n l ε ( | L ( h ) L | )
And the critical point of L ( h ) satisfies D h 1 [ L ( h ) ] O P = 0 , and L ( h ) = L .
In summary, the critical point that satisfies the assumptions (1) and (2) and Theorem 1 is the global minimum of the loss function, and Theorem 1 is proved. □

References

  1. Wang, P. Study and Application of Flatness Control Technology for Cold Rolled Strip. Master’s Thesis, Northeastern University, Shenyang, China, 2011. [Google Scholar]
  2. Dong, C.; Li, S.; Su, A. Analysis on Shape Defects of Medium and Heavy Plate and Control Measures. Tianjin Metall. 2016, S1, 27–29. [Google Scholar]
  3. Claire, N. Control of strip flatness in cold rolling. Ironand Steel Eng. 1997, 4, 42–45. [Google Scholar]
  4. Wang, C.; Zhang, Y.; Zhang, Q. Application of Effect Function in Shape Control of Cold Rolling Mill. Steel Roll. 1999, 4, 28–29. [Google Scholar]
  5. Chen, E.; Zhang, S.; Xue, T. Study on shape and gauge control system of cold strip mill. J. Plast. Eng. 2016, 23, 92–95. [Google Scholar]
  6. Liang, X.; Wang, Y.; Zhao, J.; Wang, G. Flatness Feedforward Control Models for Six High Tandem Cold Rolling Mill. Iron Steel 2009, 44, 62–66. [Google Scholar]
  7. Wang, P.F.; Yan, P.; Liu, H.M.; Zhang, D.H.; Wang, J.S. Actuator Efficiency Adaptive Flatness Control Model and Its Application in 1250 mm Reversible Cold Strip Mill. J. Iron Steel Res. Int. 2013, 20, 13–20. [Google Scholar] [CrossRef]
  8. Liu, H.; He, H.; Shan, X.; Jiang, G. Flatness Control Based on Dynamic Effective Matrix for Cold Strip Mills. Chin. J. Mech. Eng. 2009, 22, 287–295. [Google Scholar] [CrossRef]
  9. Liu, J. Research and Development of Flatness Control System in Cold Strip Rolling. Master’s Thesis, Northeastern University, Shenyang, China, 2010. [Google Scholar]
  10. Zhao, W.; Chen, F.; Huang, H.; Li, D.; Cheng, W. A New Steel Defect Detection Algorithm Based on Deep Learning. Comput. Intell. Neurosci. 2021, 2021, 1–13. [Google Scholar] [CrossRef] [PubMed]
  11. Huang, Z.; Wu, J.; Xie, F. Automatic recognition of surface defects for hot-rolled steel strip based on deep attention residual convolutional neural network. Mater. Lett. 2021, 293, 129707. [Google Scholar] [CrossRef]
  12. Roy, S.; Saini, B.S.; Chakrabarti, D.; Chakraborti, N. Mechanical properties of micro-alloyed steels studied using a evolutionary deep neural network. Mater. Manuf. Process 2020, 35, 611–624. [Google Scholar] [CrossRef]
  13. Lee, S.Y.; Tama, B.A.; Moon, S.J.; Lee, S. Steel Surface Defect Diagnostics Using Deep Convolutional Neural Network and Class Activation Map. Appl. Sci. 2019, 9, 5449. [Google Scholar] [CrossRef] [Green Version]
  14. Chun, P.J.; Yamane, T.; Izumi, S.; Kameda, T. Evaluation of Tensile Performance of Steel Members by Analysis of Corroded Steel Surface Using Deep Learning. Metals 2019, 9, 1259. [Google Scholar] [CrossRef] [Green Version]
  15. Xu, Y.; Bao, Y.; Chen, J.; Zuo, W.; Li, H. Surface fatigue crack identification in steel box girder of bridges by a deep fusion convolutional neural network based on consumer-grade camera images. Struct. Health Monit. 2019, 18, 653–674. [Google Scholar] [CrossRef]
  16. Grzegorz, P. Multi-Sensor Data Integration Using Deep Learning for Characterization of Defects in Steel Elements. Sensors 2018, 18, 292. [Google Scholar] [CrossRef] [Green Version]
  17. Fang, Z.Y.; Roy, K.; Mares, J.; Sham, C.W.; Chen, B.S.; Lim, J.B.P. Deep learning-based axial capacity prediction for cold-formed steel channel sections using Deep Belief Network. Structures 2021, 33, 2792–2802. [Google Scholar] [CrossRef]
  18. Wan, X.; Zhang, X.; Liu, L. An Improved VGG19 Transfer Learning Strip Steel Surface Defect Recognition Deep Neural Network Based on Few Samples and Imbalanced Datasets. Appl Sci. 2021, 11, 2606. [Google Scholar] [CrossRef]
  19. Wu, S.W.; Yang, J.; Cao, G.M. Prediction of the Charpy V-notch impact energy of low carbon steel using a shallow neural network and deep learning. Int. J. Min. Met. Mater. 2021; prepublish. [Google Scholar] [CrossRef]
  20. Xiao, D.; Wan, L. Remote Sensing Inversion of Saline and Alkaline Land Based on an Improved Seagull Optimization Algorithm and the Two-Hidden-Layer Extreme Learning Machine. Nat. Resour. Res. 2021, 30, 3795–3818. [Google Scholar] [CrossRef]
  21. Xiao, D.; Le, B.T.; Ha, T.T.L. Iron ore identification method using reflectance spectrometer and a deep neural network framework. Spectrochim. Acta A 2021, 248, 119168. [Google Scholar] [CrossRef] [PubMed]
  22. Bengio, Y.; Roux, N.L.; Vincent, P.; Delalleau, O.; Marcotte, P. Convex Neural Networks. In Proceedings of the International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 5–8 December 2005; pp. 123–130. [Google Scholar]
Figure 1. Comparison of loss. The horizontal axis represents epoch times (0~17.5), and the vertical axis represents loss value (0~0.8).
Figure 1. Comparison of loss. The horizontal axis represents epoch times (0~17.5), and the vertical axis represents loss value (0~0.8).
Sensors 22 00656 g001
Figure 2. Comparison of accuracy. The horizontal axis represents epoch times (0~17.5), and the vertical axis represents accuracy value (0.7~1).
Figure 2. Comparison of accuracy. The horizontal axis represents epoch times (0~17.5), and the vertical axis represents accuracy value (0.7~1).
Sensors 22 00656 g002
Figure 3. Comparison of loss. The horizontal axis represents epoch times (0~17.5), and the vertical axis represents loss value (0~0.35).
Figure 3. Comparison of loss. The horizontal axis represents epoch times (0~17.5), and the vertical axis represents loss value (0~0.35).
Sensors 22 00656 g003
Figure 4. Comparison of accuracy. The horizontal axis represents epoch times (0~17.5), and the vertical axis represents accuracy value (0.8~1).
Figure 4. Comparison of accuracy. The horizontal axis represents epoch times (0~17.5), and the vertical axis represents accuracy value (0.8~1).
Sensors 22 00656 g004
Figure 5. Reduction curve of the loss function.
Figure 5. Reduction curve of the loss function.
Sensors 22 00656 g005
Figure 6. Comparison of real and predicted values. (a) BP; (b) DNN; (c) Improved DNN. The horizontal axis represents sample points (0~175,000), and the vertical axis represents the predicted plastic shape value (18~24).
Figure 6. Comparison of real and predicted values. (a) BP; (b) DNN; (c) Improved DNN. The horizontal axis represents sample points (0~175,000), and the vertical axis represents the predicted plastic shape value (18~24).
Sensors 22 00656 g006
Figure 7. Three-dimensional fitting effect of improved DNN model.
Figure 7. Three-dimensional fitting effect of improved DNN model.
Sensors 22 00656 g007
Table 1. The common activation functions.
Table 1. The common activation functions.
Activation FunctionFormulaDescription
Sigmoid function f ( x ) = 1 1 + e x The gradient of the Sigmoid function easily falls into the saturation zone in backpropagation.
Tanh function f ( x ) = e x e x e x + e x It has soft saturation and vanishing gradient disadvantages.
ReLU function f ( x ) = { x         ,           x 0 0         ,           x < 0 As the intensity of the training increases, part of the weight update falls into the hard saturation zone, failing to update.
ELU function f ( x ) = {           x                       ,           x 0 α ( e x 1 ) ,           x < 0 ELU alleviates the vanishing gradient problem.
Swish function f ( x ) = x · sigmoid ( β x ) As the amount of data continues to increase, Swish will have better performance.
Table 2. Comparison of the best loss and accuracy value for each activation function.
Table 2. Comparison of the best loss and accuracy value for each activation function.
Activation FunctionTraining Set LossTest Set LossTraining Set AccuracyTest Set Accuracy
Improved Swish activation function0.02880.10880.99280.9852
Swish function0.02790.11590.99330.9815
ReLU function0.0290.13070.9930.982
ELU function0.03080.11330.99120.9802
Sigmoid function0.10340.10660.96580.9731
Table 3. Comparison of real and predicted values.
Table 3. Comparison of real and predicted values.
Sensor AreaPredicted ValueTrue ValueSensor AreaPredicted ValueTrue ValueSensor AreaPredicted ValueTrue Value
f939−3.299−6.014f931020.29120.623931123.51924.176
f931223.38823.659f931323.05223.336f931422.45623.036
f931521.55422.622f931619.86221.321f931717.04518.237
f931813.19814.230f93198.6479.440f93203.9024.500
f9321−1.1520.202f9322−8.183−7.844f9323−16.903−19.211
f9324−23.236−25.932f9325−25.180−26.736f9326−25.405−25.921
f9327−26.771−24.859f9328−27.216−25.265f9329−23.888−24.721
f9330−16.256−17.145f9331−6.582−6.235f9401.9622.316
f9418.5078.611f94213.30513.130f94316.82616.793
f94419.65919.976f94521.36721.737f94622.33922.387
f94722.91522.913f94823.32323.340f94923.28323.501
f941023.12423.366f941117.75117.802f9412−7.461−7.534
Table 4. Comparison of the model.
Table 4. Comparison of the model.
MSE of Training SetsMSE of Test Sets
BP7.8518.329
DNN3.2293.731
Improved DNN1.2811.305
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Liu, J.; Song, S.; Wang, J.; Balaiti, M.; Song, N.; Li, S. Flatness Prediction of Cold Rolled Strip Based on Deep Neural Network with Improved Activation Function. Sensors 2022, 22, 656. https://doi.org/10.3390/s22020656

AMA Style

Liu J, Song S, Wang J, Balaiti M, Song N, Li S. Flatness Prediction of Cold Rolled Strip Based on Deep Neural Network with Improved Activation Function. Sensors. 2022; 22(2):656. https://doi.org/10.3390/s22020656

Chicago/Turabian Style

Liu, Jingyi, Shuni Song, Jiayi Wang, Maimutimin Balaiti, Nina Song, and Sen Li. 2022. "Flatness Prediction of Cold Rolled Strip Based on Deep Neural Network with Improved Activation Function" Sensors 22, no. 2: 656. https://doi.org/10.3390/s22020656

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop