Next Article in Journal
Optimizing Sensor Position with Virtual Sensors in Human Activity Recognition System Design
Next Article in Special Issue
Post-Earthquake Building Evaluation Using UAVs: A BIM-Based Digital Twin Framework
Previous Article in Journal
A Machine Learning Framework for Balancing Training Sets of Sensor Sequential Data Streams
Previous Article in Special Issue
Trajectory Identification for Moving Loads by Multicriterial Optimization
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Temperature Prediction Using Multivariate Time Series Deep Learning in the Lining of an Electric Arc Furnace for Ferronickel Production

by
Jersson X. Leon-Medina
1,2,*,
Jaiber Camacho
3,
Camilo Gutierrez-Osorio
4,
Julián Esteban Salomón
4,
Bernardo Rueda
5,
Whilmar Vargas
5,
Jorge Sofrony
2,
Felipe Restrepo-Calle
4,
Cesar Pedraza
4 and
Diego Tibaduiza
3
1
Control, Modeling, Identification and Applications (CoDAlab), Department of Mathematics, Escola d’Enginyeria de Barcelona Est (EEBE), Campus Diagonal-Besòs (CDB), Universitat Politècnica de Catalunya (UPC), Eduard Maristany 16, 08019 Barcelona, Spain
2
Departamento de Ingeniería Mecánica y Mecatrónica, Universidad Nacional de Colombia, Cra 45 No. 26-85, Bogotá 111321, Colombia
3
Departamento de Ingeniería Eléctrica y Electrónica, Universidad Nacional de Colombia, Cra 45 No. 26-85, Bogotá 111321, Colombia
4
Departamento de Ingeniería de Sistemas e Industrial, Universidad Nacional de Colombia, Cra 45 No. 26-85, Bogotá 111321, Colombia
5
South32-Cerro Matoso S.A., Km 22 Highway SO Montelibano, Córdoba 234001, Colombia
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(20), 6894; https://doi.org/10.3390/s21206894
Submission received: 24 August 2021 / Revised: 13 October 2021 / Accepted: 13 October 2021 / Published: 18 October 2021

Abstract

:
The analysis of data from sensors in structures subjected to extreme conditions such as the ones used in smelting processes is a great decision tool that allows knowing the behavior of the structure under different operational conditions. In this industry, the furnaces and the different elements are fully instrumented, including sensors to measure variables such as temperature, pressure, level, flow, power, electrode positions, among others. From the point of view of engineering and data analytics, this quantity of data presents an opportunity to understand the operation of the system under normal conditions or to explore new ways of operation by using information from models provided by using deep learning approaches. Although some approaches have been developed with application to this industry, it is still an open research area. As a contribution, this paper presents an applied deep learning temperature prediction model for a 75 MW electric arc furnace, which is used for ferronickel production. In general, the methodology proposed considers two steps: first, a data cleaning process to increase the quality of the data, eliminating both redundant information as well as atypical and unusual data, and second, a multivariate time series deep learning model to predict the temperatures in the furnace lining. The developed deep learning model is a sequential one based on GRU (gated recurrent unit) layer plus a dense layer. The GRU + Dense model achieved an average root mean square error (RMSE) of 1.19 °C in the test set of 16 different thermocouples radially distributed on the furnace.

1. Introduction

Structural health monitoring (SHM) remains a priority in large-scale industrial applications because of the multiple advantages in its implementation such as the use of sensors permanently attached to the structure for continuous monitoring and its readily available form for the application of data-driven approaches to determine the health of the structure under evaluation [1,2]. Structures in industrial processes require special attention since the inability to do so may put at risk its operation. Structural health uncertainty, in conjunction with variations in the operational or environmental conditions, increases the risk of accidents, which in turn may result in catastrophic events [3]. One example of structures subjected to extreme operational conditions is the furnaces used in the smelting industry where temperatures, pressures, flows, among other variables, can vary considerably due to changes in the inputs (e.g., chemical composition of the raw material) or the operational conditions. This implies that operators are obliged to continuously monitor the structural health of the furnace before applying any changes in the operation set points, in order to maintain conditions that allow the secure operation of the system and the health of the structure. SHM is of particular interest for electric arc furnaces (EAF), which are characterized by the action of heating the materials using a covered electric arc for the smelting process. Routine operation causes wear in the wall lining of the furnace, hence monitoring wall thickness is of special interest to avoid run-outs of the smelting material. One strategy to evaluate the state of the structure is to directly monitor the wall thickness via routine inspection using specialized techniques. However, it is an expensive and difficult task because of the scale of the system. For this reason, it is of interest to use other measurements that can indirectly account for the thickness of the wall lining. As it turns out, the measured temperature in the wall is a good indicator of its health and can be used for monitoring tasks in a continuous (on-line) way.
Some recent works that use machine learning techniques to predict temperature variables within the smelting process have addressed the problem from different points of view. Mishra et al. [4] compared five deep learning models for multivariate prediction of time series temperatures. The study yielded as a result that a deep convolutional network (DCN) performs best with wavelet and fast Fourier transform (FFT). An online estimation of electric arc furnace tap temperature by using fuzzy neural networks was developed by Fernandez et al. [5] and its application of this helped reduce the consumption of energy in an electric arc furnace. In the work of Fontes et al. [6], the hot metal temperature in a blast furnace was predicted using an approach based on fuzzy c-means (FCM) and exogenous nonlinear autoregressive model (NARX); the estimate was later implemented as a soft sensor for predicting temperature. In Shockaert and Hoyez [7] a multivariate time series approach was built with a deep generative CycleGAN model combined with a long short-term memory (LSTM)-based autoencoder (AE). Particularly, that approach handles a transfer learning methodology in which data obtained from a source furnace is used to train a model that can be used to evaluate data from a target furnace. The forecasting of the hot metal temperature in a blast furnace is shown in the work of Iffat et al. [8]. Specifically, an optimal time lag at which the input variables have an impact on the hot metal temperature is determined. Additionally, an incremental learning methodology that considers changes in raw material composition, process control methods, and aging equipment was developed.
In order to improve the efficiency in a blast furnace, a self-organizing Kohonen neural network approach was developed in [9]. This approach managed to complement the control of process operating parameters for the blast furnace process. An artificial neural network (ANN) model was applied to predict the slag and metal composition in a ferromanganese production unit with a submerged arc furnace [10]. The advantages of this ANN included the reduction of power and coke consumption. Similarly, in the work of Ducic et al. [11], an ANN was derived as an intelligent soft sensor in the process of white cast iron production. An increase in productivity, as well as material and energy efficiency, can be translated into a reduction of the environmental impact and cost of the steel-making process in a basic oxygen furnace (BOF); in [12], the aforementioned objectives were reached through the use of standard machine learning models to predict the end-point targets of variables like the final melt temperature and upper limits of the carbon and phosphorus content with minimum material loss. A random forest algorithm was used in [13] to improve the quality of steel casting for tire reinforcement; 140 process variables were used as features and the output can take values 0 or 1 depending on whether the casting was rejected or not. The best area under the receiver operating characteristics (AUROC) in the test set was 0.85 obtained by the random forest classification method. The prediction issue of the amount of alloying additives in order to obtain the desired chemical composition of white cast iron was solved applying a neural network model in [14]. A three-month-long monitoring of the metal melting process data set was used. Besides the data was split into training and test sets founding that the neural network model reached a mean squared error of 3.31% in the test set.
As previous works have shown, different variables can be predicted by using historical operation data to evaluate the health of furnaces, or to predict the behavior of the process when inputs are changed. Therefore, the development of prediction models becomes one of the main necessities in the areas of operation analysis, control, and maintenance of EAFs and remains an open research area. As a contribution, this work presents the development of a deep learning model to predict the temperature in the lining of the EAF. Details about the experimental setup, the data acquisition from sensors, its preprocessing step, model development, and its validation are also presented in this work. This work was carried out in a joint effort between academia and industry, Universidad Nacional de Colombia and Cerro Matoso S.A. (CMSA). The interested reader can acquire some more background of the research process by reviewing some of the previous works developed by the authors, where the problems associated with sensor networks and continuous monitoring in this kind of furnace, including temperature monitoring [15,16], gap monitoring [17], and thickness monitoring [18] using ultrasonic and ground penetrating radar (GPR) methods are also tackled.
This paper is organized as follows: Section 2 includes a theoretical background where some concepts about the company, the process, and the methods used in the methodology are briefly introduced. Section 3 provides information about the developed methodology including a description of the dataset used for the validation. Section 4 presents the results and discussion; finally, conclusions are included in the last section.

2. Theoretical Background and Related Works

2.1. Cerro Matoso S.A.

As context to the work presented herein, Cerro Matoso S.A. (CMSA) is one of the world’s major producers of ferronickel and it is operated by South32. This is an open-cut mine operation in northern Colombia, with more than 35 years of operation in the region. More details about the process developed by CMSA can be found directly on its web page [19]. A brief description of nickel and the ferronickel metallurgical process are included next in order to contextualize the development of the methodology.

2.2. Nickel and Ferronickel Metallurgical Process Description

Nickel is metal and is commonly used in stainless steel production. Stainless steel that includes nickel is used in the food processing, transportation, and manufacturing industries because of its advantages such as being heat-resistant, resistant to damage, and easy to keep clean. It is commonly used in everyday household items as well [19].
In the case of CMSA, the main product is ferronickel (FeNi), which is a material used in different industries such as electronics, manufacturing, and automotive, among others. This material is obtaining in a metallurgical procedure after mining laterites and high-Ni sulfide ores [20]. Figure 1 shows a diagram depicting the ferronickel extraction process. This process starts with the ore extraction in an open-cut mine, which is tested and classified. The stored ore is dried in a calcine furnace (kiln furnace) and afterward processed in an electric arc furnace where the smelting process is performed. Figure 2 shows a panoramic of a section of the furnace. Although all parts of the process are important, the monitoring of the furnace is a vital task because of the risks associated with its operation. It is necessary to remark that this element works 24/7, and the aim of the monitoring system is to reduce the number of maintenance activities by an enhanced operation of the furnace. The experience of the furnace operators, who are making decisions based on the data from the process every day, is one of the main sources of knowledge that can make this happen, but for they this they require good online information and predictions.

2.3. Deep Learning Predictive Methods

As is shown in Figure 1, the ferronickel extraction process requires several steps and involves multiple inputs and outputs. This complexity requires the use of advanced strategies for data analysis and it is here where neural networks present solutions to tackle problems that involve sequential processing of data [21]. The goal of this paper is to develop a predictor of the temperature behavior in the furnace using mainly data obtained from thermocouple sensors located at the lining furnace. One of the requirements of the proposed method is to effectively model long-term dependencies between variables, therefore it is necessary to use information from earlier time windows (past information) in order to be able to accurately predict the temperature (future predictions). This highlights the fact that it is necessary to develop models that can handle variable-length input sequences, that are able to track long-term dependencies in the data, that can maintain information about the order of the sequences, and share parameters across the entirety of the sequence.
Feedforward neural networks are not able to maintain information about a previous event in a sequence of events [22]. In contrast, recurrent neural networks (RNN) have loops in their architecture, which allows for information to persist over time. These networks are called recurrent because the information is being passed from one time-step to the next internally within the network [23].
The RNNs use a training algorithm called backpropagation through time [24]. Errors are backpropagated at each time step, and then, finally across all time steps all the way from where we are currently to the beginning of the sequence [25]. This is the reason why it is called backpropagation through time. The computation of the gradient, that is the derivative of the loss with respect to the parameters tracing all the way back to the initial state, requires many repeated multiplications of the weight matrix as well as repeated use of the derivative of the activation function. In some cases, this is a problem because gradients are too small. This problem is well-known as the vanishing gradient problem [26]. There are three different ways to overtake the vanishing gradient problem: (i) choosing the activation function, (ii) initializing the weights cleverly (close to the optimal solution), and (iii) designing the network architecture to actually be able to handle this efficiently. This work focuses on the use of the latter solution, which uses a slightly more complex recurrent unit that can track long-term dependencies in the data more effectively by controlling what information is passed through and what information is used to update its internal state, i.e., the gated cell. Two types of gated cells are described next, namely the long short-term memory (LSTM) and the gated recurrent unit (GRU).

2.3.1. Convolutional Neural Networks

Convolutional neural networks (CNNs) are feedforward artificial neural networks that use the convolution operation instead of matrix multiplication. The main aspects of the convolution layer in CNNs are its sparse local connectivity and filters, which significantly reduce the number of network parameters, while simultaneously increasing its performance [27]. Features in CNN are not hand engineered, but learned; this property reduces the preprocessing stage.

2.3.2. Long Short-Term Memory Network

In an LSTM network, the repeating unit contains different interacting layers. These layers interact to selectively control the flow of information within the cell. This enables LSTM to track and store information throughout many time steps. The key building block behind the LSTM is the gate, which functions to enable the LSTM to selectively add or remove information to its cell state. LSTM processes information through four simple steps: forget, store, update, and output. These networks must first forget irrelevant history, then perform the computation to store relevant parts of new information, use these two steps together to selectively update their internal state, and finally, generate an output. The internal description of an LSTM unit is illustrated in Figure 3. LSTM considers an input sequence { x 1 , x 2 , , x t } at time t. The output gate determines the new state h t , where t is the time step. The following equations describe the internal operations carried out in an LSTM internal unit [28].
f t = σ W f × h t 1 , x t + b f
i t = σ W i × h t 1 , x t + b i
o t = σ W o × h t 1 , x t + b o
C ˜ t = t a n h W c × h t 1 , x t + b c
C t = f t × C t 1 + i t × C ˜ t
h t = o t × tanh C t
where, C t 1 , C ˜ t , and C t are unit memory, W c , W o , W f , and W i are weight matrices; b c , b o , b f and b i are bias vectors.

2.3.3. Gated Recurrent Unit

The GRU [29] is a modified version of the LSTM cell. It combines long and short-term memory into its hidden state. The GRU has two gates, on one hand, the update gate and on the other hand the reset gate. These gates allow maintaining a balance between the information to retain and forget. The outputs of the GRU unit are z t (see Equation (7)) and r t (see Equation (8)).
The internal description of a GRU unit is depicted in Figure 4. The following equations describe the internal operations in the GRU unit [28].
z t = σ W z × h t 1 , x t + b z
r t = σ W r × h t 1 , x t + b r
m t = tanh W s × r t 1 , x t + b s
h t = 1 z t × h t 1 + z t + m t
where, W r , W z and W s are weight matrices; and b r , b z , and b s are bias vectors.

3. Temperature Prediction Methodology for the Furnace Lining

This section is devoted to introducing the proposed methodology for temperature prediction in the wall of the ferronickel furnace studied. Two general steps are considered after the data acquisition step: the data cleaning process and the development of the deep learning model as depicted in Figure 5. Although the methodology considers some particular elements and variables of this specific smelting process, it can be generalized to other complex processes where a big number of sensors are used and it is necessary to predict the behavior of a variable.
Before presenting each step of the methodology, some context with regards to the data set obtained from the data acquisition system is given next.

3.1. Dataset for Methodology Validation

The data set corresponds to the measured variables of an electric arc furnace for ferronickel production located at Cerro Matoso SA (South32 company). The data set used was sampled every 15 min for a period of 416 days between the 11 August 2018 and 30 September 2019. The data set is composed of a total of 40,000 instances. Regarding the attributes, also called variables or features, an in-depth analysis with a group of expert furnace operators defined a group of 49 variables selected due to their importance in the furnace operation. These 49 variables are detailed in Table 1 and serve as input variables to train and test the developed multivariate time series temperature prediction system.
The electric arc furnace is built with the integration of 72 panels radially distributed on its perimeter. Each panel is composed of four plate coolers; these four plate coolers are located at four different heights labeled Level 1, Level 2, Level 3, and Level 4 as depicted in Figure 6. Thus, there are 72 × 4 = 288 plate coolers radially distributed along the furnace. Each plate cooler has a thermocouple for temperature lining monitoring. Figure 6 shows a section of the wall of the furnace where it is possible to observe the refrigeration system composed of 4 levels of plate coolers per panel, where the thermocouples, whose temperature measurement must be predicted, are located.
Due to the high number of plate cooler thermocouples in the lining furnace, a discrete group of 16 thermocouples were selected in this work as output variables to be predicted. Four panels of the furnace belonging to the North-West (NW), South-West (SW), South-East (SE), and North-East (NE) quadrants were selected. Each panel has four thermocouples, thus a 4 × 4 = 16 thermocouples in total were selected. The distribution of these 16 thermocouples is illustrated in Figure 7. The location of the three furnace electrodes (E1, E2, and E3) is detailed in Figure 7 (left).

3.2. Data Pre-Processing Step

The ferronickel production process carried out at the CMSA facilities is made up of a large number of variables. Variables are acquired from different sources, including multiple on-line sensors, and collected by a data acquisition system (DAQ). The collected data can contain errors because of failures in sensors, noise, or missed data that the system fails to capture or store. These errors require revision and errors must be eliminated in the pre-processing step in order to reduce the errors in the model to be developed. Originally 1180 variables were provided by CMSA. Together with experts from the process operations area, some considerations about the range of the variables were identified, and the elements shown in Figure 8 were defined as the set of rules to be considered. These elements constitute a workflow with seven steps, which contains different types of problems that might be present in the initial data set, for example, strings of characters in numeric variables, negative temperatures, variables that remain in a single value, variables with null data. With this workflow, it was possible to find variables that consistently presented problems in the data, and for this reason, they were eliminated. It is also noted that given the large number of variables available, it was not necessary to carry out the data restoration processes, avoiding in this way the appearance of gaps in the data set. The results obtained after applying the proposed workflow allowed to debug 340 variables with errors, which indicates that 28% of the data has problems.
The following steps in the data cleaning process were considered [15]:
  • High variance: Data with values outside of the operational range should not be considered. Thus, a univariate measure for the measurement of quality based on percentage changes is calculated. A variable representation with variance is shown in Figure 9a.
  • Strings: the variables that had non-numeric values were encoded with numerical values (see Figure 9b).
  • Remove duplicates: The identification of duplicate variables results in drop the duplicate and stay with a single variable (see Figure 10a).
  • Empty and null values: A 2% was selected as threshold, thus, variables with more than 98% with empty or null values are dropped. (see Figure 10b)
  • Zero variance: If more than 50 % of the data of a variable does not vary and remains in a constant value the variable is dropped (Figure 11a).
  • Unique values: Because it is desirable to find relationships between variables those that remain constant over time are dropped (see Figure 11b).
  • Negative temperatures: The normal operating range of variables that measure temperature always takes positive values. Due to the above, some variables with negative temperatures are identified and these are dropped (Figure 11c).
The amount of variables eliminated in each of the categories in the data cleaning process is described in Table 2. As a result, after performing the cleaning process, 840 variables were obtained in a cleaned data set. Subsequently, through suggestions made by the furnace operators and the judgment of expert engineers belonging to CMSA, the 49 variables of the temperature prediction model developed in this work were selected. These 49 variables are listed in Table 1.

3.3. Development of the Multivariate Time Series Deep Learning Model

The following section will the steps carried out in order to achieve our goal of training a temperature prediction model. To this end, we will first discuss the steps taken in order to define a suitable data set. This is followed by a discussion on the definition and development of the RNN proposed to be used in this paper.
The programming language selected to implement the temperature prediction models is Python. Together with this programming language the following libraries are used for data management, neural network training and visualization, among other functions:
  • Data manipulation: Pandas and NumPy
  • Data visualization: Matplotlib and Seaborn
  • Neural networks training: scikit-learn [30], TensorFlow [31], and Keras [32].

3.3.1. Definition of the Dataset

The data set has input variables of different magnitudes and values. So the different values are scaled so that they are in the interval between −1 and 1. The target data comes from the same data set as the input signals, because they are the output thermocouple data that simply shifts in time.
The number of time steps that it will shift the target data is predefined. The dataset was sampled to have one observation every 15 min, thus there are 96 observations over 24 h. In particular, the shift is used to predict temperatures two hours in the future.
Due to the large number of instances that were taken in the dataset (40,000) it is impractical to perform a conventional division of 70% of the data for training and the remaining 30% for testing. In contrast, and due to the desire to have the greatest amount of data for training, the decision was made to divide the data into 90% for training and the remaining 10% for testing. In addition, the input and output variables for the training and test sets were defined. The dataset must be prepared as two-dimensional NumPy arrays. In this case, there are 49 input signals and 16 output signals.
Instead of training the recurrent neural network on the entire sequence of 40,000 observations, a function is used to create a batch of shorter 250 subsequences randomly selected from the training data. Thus, every sequence had a size of =1152 steps corresponding to 12 days. This period of time is defined because it is in which the pile of calcined material in the furnace is changed. The 40,000 data used in the training and testing of the temperature prediction model were cleaned before the development of the neural network. Thus, there are 250 random sequences (batch) with a size of 1152 steps and all the data belonging to this batch is clean following the steps described in Section 3.2, where it was determined as a data cleaning policy that if a variable had any of the problems indicated, it would be eliminated, this was possible given a large amount of information available and variables in the initial data set. For that reason, the variables with data problems were not available for the training data set, because their information did not represent the real behavior of the furnace, therefore it is clarified that there was no partial elimination of values that generated discontinuity in the time series.
As described, the training is performed by taking 250 random sequences. However, to carry out the testing, the complete sequence of data is taken in the test set corresponding to 4000 records. These 4000 records in the test set are sufficient and still providing a good enough estimate of the model performance. In addition, the model performance is monitored after each epoch on the test set and only if the performance is improved on the test set the weights of the recurrent neural network are saved for the next epoch.

3.3.2. Creation of the Recurrent Neural Network

The neural network and its different layers are created using TensorFlow in a sequential model. The first layer consists of the use of a cell type gated recurrent unit (GRU) to create a recurrent neural network. This GRU layer had 250 outputs for each time step in the sequence. The information that enters this first GRU layer of the model is a batch of sequences of arbitrary length where each observation has several input signals. The GRU network generates a batch of sequences of 250 values. You want to predict 16 output signals; thus, a dense layer is added in the deep learning model that assigns 250 values to only 16 output values corresponding to the 16 thermocouples that it is desired to predict. A sigmoid activation function is used to ensure that the values are within the normalized values.
After each epoch of the neural network, the performance of the model in the test set is monitored and the model weights are only saved if the performance improves in the test set. In the training process, a batch of short sequences is randomly selected from the total training data. In this case the training data has 36,000 instances. In contrast, for the validation data, the entire sequence is run from the 4000 instances in the test set and the prediction accuracy is measured on that entire sequence.
It is important to discuss the loss function, learning rate, and warmup period. The loss function that is minimized is the mean square error (MSE). A warmup period of 50 time steps is assigned to the model so that the precision of these first 50 steps is not used in the stall function. The inclusion of this warmup period allows the model to present a better behavior for each of the 16 outputs in terms of root mean square error RMSE. Adam [33] was selected as an optimizer and an initial learning rate of 1 × 10 3 is used. If the loss of validation has not improved since the last epoch, the learning rate changes to 1 × 10 4 . A two-layer model was defined, one GRU and the other dense. The output form of (None, None, 16) shown in Figure 12 means that the model will generate a batch with an arbitrary number of sequences, each of which has an arbitrary number of observations, and each observation has 16 output signals.

3.3.3. Training of the Recurrent Neural Network

A single “epoch” does not correspond to a single training set processing, due to how the batch generator randomly selects subsequences from the training set. Instead, “steps-per-epoch” is selected to have an epoch processed in a few minutes. In this case, the number of steps-per-epoch used is equal to 100. The parameters of the joint GRU + Dense model are described in Table 3.

3.3.4. Performance of the On-Line Prediction Model

The final step is to compare the predicted and true output signals. The time series prediction model performance is calculated using the root mean squared error (RMSE):
R M S E = 1 M i = 1 M ( y ˜ i y i ) 2
where M is the number of data points in the time series to be estimated, y i is the actual value of time series, and y ˜ i is the estimated value at time i by the prediction model [34].
The steps to develop the different deep learning models to predict the temperature are described in this section. These steps are illustrated in Figure 13.

4. Results and Discussion

As a result of the multivariate time series deep learning model evaluation, a comparison of the predicted and true behaviors for one thermocouple in the train and test sets is shown in Figure 14. It can be deduced the similarity between the true and predicted values in both cases.

4.1. RMSE Study in the Train and Test Sets for Six Different Deep Learning Models

A comparative study showing the 16 thermocouple results in the test set is shown in Figure 15. In each subfigure, the true value is depicted in blue and the predicted value is depicted in orange. From Figure 15 is evident that Level 1 reaches the lowest temperatures in general independently of the quadrant. In contrast, the highest temperatures in the furnace were reached in Level 4 due to its bottom location in the furnace. When a comparison between quadrants is performed, it is evident that the NW and SW sectors exhibit higher values than SE and NE sectors in the furnace. This is probably due to an erosion in the refractory lining in the NW and SW quadrants of the furnace.
Table 4 shows the RMSE values for each one of the 16 thermocouples for the train and test sets. A comparison study of the RMSE behavior was carried out with six different deep learning models. These deep learning models were:
  • GRU (250 Cells) + Dense
  • GRU (128 cells) + GRU (64 cells) + Dense
  • LSTM (250 cells) + Dense
  • GRU (128 cells) + LSTM (128 cells) + Dense
  • CONV1D (128 cells) + Dense
  • CONV1D (128 cells) + LSTM (128 cells) + Dense
As shown in Table 3 the best deep learning model was the GRU (250 Cells) + Dense. It may be observed that the best average RMSE value of 1.19 in the test set was obtained by the GRU (250 Cells) + Dense model. The CONV1D layer automatically extracts features from the input time series during training. As can be observed in Table 3 the second-best model was the CONV1D (128 cells) + Dense reaching an average RMSE value of 1.48 in the test set. This behavior was obtained by using one-dimensional filters to capture the temporal properties in the CONV1D layer to describe the temporal pattern of the input series [35]. The thermocouples in Level 1 NE and Level 2 NE reached the lowest RMSE values. In this case, the two aforementioned thermocouples belong to the NE plate cooler and correspond to the upper and middle Levels 1 and 2 showing their high correlation. In general, the NE quadrant exhibits the best RMSE values and it is evident that the NE quadrant presents the lower temperatures of the entire furnace.
Particularly, the thermocouple in Level 4 NE reached a maximum temperature of 70 °C. In contrast, the thermocouples in Level 4 NW and SW reached temperatures above 120 °C. The worst RMSE values were obtained for the thermocouples of Level 4 NW and Level 4 SW, with values of 2.94 and 2.61 respectively in the test set. This behavior is mainly due to the influence of the electrodes at this position in the furnace area.
The loss behavior of the training and test sets is shown in Figure 16. A change in behavior is evident when the model reaches epoch 8 since, from then on, the decrease in the loss presents less intensity. Due to the small values of the loss in magnitude a logarithmic y-axis is used. It is evident that the values of the training set are lower than those of the test set as the epochs increase. Additionally, the observed behavior allows us to obtain a model while preventing over-fitting.

4.2. Robustness Study against the Random Sampling in the Sequences in the Training Set

Table 5 shows the average RMSE behavior for train and test sets after performing five iterations for the GRU (250 Cells) + Dense model. All the parameters remained constant in every iteration. The slight changes are due to the random sampling in selecting 100 sequences to train. From Table 5 it is evident that the RMSE remains in the same order of values through the iterations. Therefore, the robustness of the model can be evidenced.

4.3. Study of the Variation of the Number of Sequences in the GRU Model

A next study changing the size of sequences that are part of the batch in the training is performed to evaluate the RMSE behavior of the GRU (250 Cells) + Dense model. The variation results of the size of sequences using 10, 50, 100, 150, 200, and 250 random sequences in the training is depicted in Table 6. The best average RMSE value in the test set was achieved when using 250 random sequences.

4.4. Moving Origin Four-Fold Cross Validation Strategy in a Time Series Approach

A four-fold moving origin [36] cross-validation strategy in a time series approach was performed. The moving origin four-fold cross-validation strategy is depicted in Figure 17. This cross-validation strategy is carried out by changing the percentage of data in the train and test sets in different iterations. Due to the total number of instances is 40,000 a four-fold split was selected. As a result, four different folds were evaluated. The size of the test set in each fold is equal to 8000 data. The results of the average RMSE values obtained in the train and test sets in each fold are shown in Table 7. It is evident the low RMSE value of 0.52 in the train set for fold #1. As the iterations are performed and the amount of data in the train set increases, the average RMSE also increases in the train set. In contrast, a decreasing average RMSE behavior across the folds is evident in the test set. This is a consequence of the increased data in the train set, which allows for better modeling of the furnace behavior.

4.5. Sensitiveness against Training/Test Ratio

A study of the sensitiveness against training/test ratios of 90/10, 85/15 and 80/20 was performed on the entire 40,000 data. The results of average RMSE are shown in Table 8. An average RMSE value remained constant for the train set in all the cases. In contrast, the behavior of the average RMSE increased in the test set as the size of the test set increased from 10% to 20%. The best average RMSE of 1.19 for the test set was obtained for the 10% configuration, this indicates that the more data that belongs to the training set, the better the model will be able to generalize the behavior of the 16 output thermocouples.

4.6. Influence of the Variables Used in the Model

To determine the behavior of the model in relation to the input variables, the GRU model was trained by eliminating one variable at a time. Figure 18 shows the results of RMSE for this analysis in the training set. A low influence of some of the modifications is observed, for example, the omission of inputs related to the furnace power. On the other hand, the variables related to the electrodes arc have a considerable influence on the decrease of errors in training.
Figure 19 shows the results of the model’s test set. Remarkable changes are observed in relation to the test results. The elimination of the variables related to the position of the electrodes (being the variable that decreases the RMSE the most) is the omission that has the greatest influence on the increase in this error. The complexity of the developed prediction model can be simplified by reducing the number of inputs, specifically those related to the position of the electrodes, without affecting the results in the training and testing processes. For example, it can be observed that the variables electrode position and calcine supply are not improving the model performance. However, it can be observed that current variables are affecting the RMSE when they are removed, hence they are important to the model.

5. Conclusions

In this paper, a multivariate time series deep learning model was developed to predict the temperature behavior in an electric arc furnace. The developed temperature prediction methodology was tested on a dataset of 416 days of an electric arc furnace operation, corresponding to 40,000 instances. Sixteen thermocouples radially distributed in the furnace at four different height levels were selected as output variables. The results yielded by the GRU (250 Cells) + Dense deep learning model showed an average RMSE of 1.19 °C for the test set using a training/test ratio of 90/10. This shows the goodness of the prediction in the SHM system for furnace lining temperature monitoring.
It was found that approximately 28% of the original dataset presented abnormalities, thus, it was very important to carry out a data preprocessing step including data cleansing, outlier removal, and removing redundant, null, and unwanted values.
The developed deep learning model allowed us to perform temperature predictions in the lining of the furnace at 2 h in the future. Consequently, the predicted behavior of the furnace facilitates decision-making associated with the possible high temperatures of the furnace hearth due to changes in the operational variables. These predictions contribute to carrying out correct structural health monitoring and preventive control of the furnace lining erosion, caused by excess temperature.
This research allowed us to determine which variables are relevant in the prediction of temperature, confirming the hypothesis about the relationship between each variable and the furnace lining temperature. This was necessary to determine the input variables of the multivariate time series deep learning model.
In future work, an attention-based model inspired by an encoder-decoder approach will be applied to predict the furnace temperature considering relations between variables in long and short term periods of time. Moreover, a more sophisticated architecture for the CNN1D, involving several layers of convolution and maxpooling will be proved in order to identify their capacity to catch more abstract features. Besides, the developed model will be tested in another electric arc furnace, and their ability to predict its lining temperature will be compared.

Author Contributions

All authors contributed to the development of this work, specifically their contributions are as follow: conceptualization, D.T., C.P., F.R.-C. and J.S.; data organization and pre-processing, J.E.S. and J.C.; methodology, J.X.L.-M., C.G.-O., D.T., C.P., F.R.-C. and J.S.; validation, J.X.L.-M., W.V. and B.R. All authors have read and agreed to the published version of the manuscript.

Funding

This work has been funded by the Colombian Ministry of Science through the grant number 786, “Convocatoria para el registro de proyectos que aspiran a obtener beneficios tributarios por inversión en CTeI”.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Acknowledgments

The authors would like to thank the Editor and the anonymous reviewers for their valuable comments and suggestions. The authors express their gratitude to the CMSA team, especially to Luis Bonilla and the structural control team for providing all data used in this work. In the same manner to Janneth Ruiz, Cindy Lopez, and Carlos Galeano Urueña for their support along the development of this work.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
LSTMLong Short Term Memory
GRUGated Recurrent Unit
CNNConvolutional Neural Network
RNNRecurrent Neural Network
DCSDistributed Control System

References

  1. Anaya, M. Design and Validation of a Structural Health Monitoring System Based on Bio-Inspired Algorithms. Ph.D. Thesis, Universitat Politècnica de Catalunya, Barcelona, Spain, 2016. [Google Scholar]
  2. Tibaduiza, D.A.; Torres-Arredondo, M.A.; Mujica, L.; Rodellar, J.; Fritzen, C.P. A study of two unsupervised data driven statistical methodologies for detecting and classifying damages in structural health monitoring. Mech. Syst. Signal Process. 2013, 41, 467–484. [Google Scholar] [CrossRef]
  3. Tibaduiza, D. Design and validation of a structural health monitoring system for aeronautical structures. Ph.D. Thesis, Universitat Politècnica de Catalunya, Barcelona, Spain, 2013. [Google Scholar]
  4. Mishra, S.; Bordin, C.; Taharaguchi, K.; Palu, I. Comparison of deep learning models for multivariate prediction of time series wind power generation and temperature. Energy Rep. 2020, 6, 273–286. [Google Scholar] [CrossRef]
  5. Fernández, J.M.M.; Cabal, V.Á.; Montequin, V.R.; Balsera, J.V. Online estimation of electric arc furnace tap temperature by using fuzzy neural networks. Eng. Appl. Artif. Intell. 2008, 21, 1001–1012. [Google Scholar] [CrossRef]
  6. Fontes, D.O.L.; Vasconcelos, L.G.S.; Brito, R.P. Blast furnace hot metal temperature and silicon content prediction using soft sensor based on fuzzy C-means and exogenous nonlinear autoregressive models. Comput. Chem. Eng. 2020, 141, 107028. [Google Scholar] [CrossRef]
  7. Schockaert, C.; Hoyez, H. MTS-CycleGAN: An Adversarial-based Deep Mapping Learning Network for Multivariate Time Series Domain Adaptation Applied to the Ironmaking Industry. arXiv 2020, arXiv:2007.07518. [Google Scholar]
  8. Iffat, U.; Bhatia, S.; Tantar, A.; Sanz, J.; Schockaert, C.; Schimtz, A.; Giroldini, F.; Reuter, Y.; Hansen, F. New Digital Services for Manufacturing Industry Using Analytics: The Case of Blast Furnace Thermal Regulation. In Proceedings of the 2018 IEEE 20th Conference on Business Informatics (CBI), Vienna, Austria, 11–14 July 2018; Volume 2, pp. 89–91. [Google Scholar]
  9. Shnayder, D.; Barbasova, T.; Lapteva, Y. Enhancing Blast Furnace Control Efficiency Based on Self-Organizing Kohonen Neural Networks. In Proceedings of the 2018 International Conference on Industrial Engineering, Applications and Manufacturing (ICIEAM), Moscow, Russia, 15–18 May 2018; pp. 1–4. [Google Scholar]
  10. Singh, V.; Biswas, A.; Tripathy, S.; Chatterjee, S.; Chakerborthy, T. Smart ore blending methodology for ferromanganese production process. Ironmak. Steelmak. 2016, 43, 481–487. [Google Scholar] [CrossRef]
  11. Dučić, N.; Ćojbašić, Ž.; Slavković, R.; Jordović, B.; Purenović, J. Optimization of chemical composition in the manufacturing process of flotation balls based on intelligent soft sensing. Hem. Ind. 2016, 70, 603–612. [Google Scholar] [CrossRef] [Green Version]
  12. Bae, J.; Li, Y.; HL, N.S.; Mathiason, G.; Kojola, N. Using Machine Learning for Robust Target Prediction in a Basic Oxygen Furnace System. Metall. Mater. Trans. Process. Metall. Mater. Process. Sci. 2020, 51, 1632–1645. [Google Scholar] [CrossRef]
  13. Cuartas, M.; Ruiz, E.; Ferreño, D.; Setién, J.; Arroyo, V.; Gutiérrez-Solana, F. Machine learning algorithms for the prediction of non-metallic inclusions in steel wires for tire reinforcement. J. Intell. Manuf. 2021, 32, 1739–1751. [Google Scholar] [CrossRef]
  14. Dučić, N.; Jovičić, A.; Manasijević, S.; Radiša, R.; Ćojbašić, Ž.; Savković, B. Application of Machine Learning in the Control of Metal Melting Production Process. Appl. Sci. 2020, 10, 6048. [Google Scholar] [CrossRef]
  15. Camacho-Olarte, J.; Torres, J.E.S.; Jimenez, D.A.G.; Medina, J.X.L.; Vargas, R.C.G.; Cardenas, D.A.V.; Gutierrez-Osorio, C.; Rueda, B.; Vargas, W.; Burgos, D.A.T.; et al. A Data Cleaning Approach for a Structural Health Monitoring System in a 75 MW Electric Arc Ferronickel Furnace. Eng. Proc. 2020, 2, 21. [Google Scholar] [CrossRef]
  16. Leon-Medina, J.X.; Vargas, R.C.G.; Gutierrez-Osorio, C.; Jimenez, D.A.G.; Cardenas, D.A.V.; Torres, J.E.S.; Camacho-Olarte, J.; Rueda, B.; Vargas, W.; Esmeral, J.S.; et al. Deep Learning for the Prediction of Temperature Time Series in the Lining of an Electric Arc Furnace for Structural Health Monitoring at Cerro Matoso (CMSA). Eng. Proc. 2020, 2, 23. [Google Scholar] [CrossRef]
  17. Tibaduiza, D.A.; León, J.; Bonilla, L.; Rueda, B.; Zurita, O.; Forero, J.C.; Vitola, J.; Segura, D.; Forero, E.; Anaya, M. Gap Monitoring In Refurbishment Tasks In A Ferronickel Furnace At Cerro Matoso SA. In Proceedings of the XI International Conference on Structural Dynamics-EURODYN 2020, Athens, Greece, 22–24 June 2020; pp. 4722–4729. [Google Scholar] [CrossRef]
  18. Tibaduiza, D.A.; Leon-Medina, J.X.; Gomez, R.; Ricardo, J.; Rueda, B.; Zurita, O.; Forero, J.C. Structural Health Monitoring System for Furnace Refractory Wall Thickness Measurements at Cerro Matoso SA. In European Workshop on Structural Health Monitoring; Rizzo, P., Milazzo, A., Eds.; Springer International Publishing: Cham, Switzerland, 2021; pp. 414–423. [Google Scholar]
  19. Cerro Matoso–South32. 2021. Available online: https://www.cerromatoso.com.co/ (accessed on 15 July 2021).
  20. Keskinkilic, E. Nickel laterite smelting processes and some examples of recent possible modifications to the conventional route. Metals 2019, 9, 974. [Google Scholar] [CrossRef] [Green Version]
  21. Najafabadi, M.M.; Villanustre, F.; Khoshgoftaar, T.M.; Seliya, N.; Wald, R.; Muharemagic, E. Deep learning applications and challenges in big data analytics. J. Big Data 2015, 2, 1–21. [Google Scholar] [CrossRef] [Green Version]
  22. Makris, D.; Kaliakatsos-Papakostas, M.; Karydis, I.; Kermanidis, K.L. Combining LSTM and feed forward neural networks for conditional rhythm composition. In International Conference on Engineering Applications of Neural Networks; Springer: Cham, Switzerland, 2017; pp. 570–582. [Google Scholar]
  23. Schuster, M.; Paliwal, K.K. Bidirectional recurrent neural networks. IEEE Trans. Signal Process. 1997, 45, 2673–2681. [Google Scholar] [CrossRef] [Green Version]
  24. Werbos, P.J. Backpropagation through time: What it does and how to do it. Proc. IEEE 1990, 78, 1550–1560. [Google Scholar] [CrossRef] [Green Version]
  25. Lim, B.; Zohren, S. Time-series forecasting with deep learning: A survey. Philos. Trans. R. Soc. 2021, 379, 20200209. [Google Scholar] [CrossRef]
  26. Hochreiter, S. The vanishing gradient problem during learning recurrent neural nets and problem solutions. Int. J. Uncertain. Fuzziness -Knowl.-Based Syst. 1998, 6, 107–116. [Google Scholar] [CrossRef] [Green Version]
  27. Puruncajas, B.; Vidal, Y.; Tutivén, C. Vibration-Response-Only Structural Health Monitoring for Offshore Wind Turbine Jacket Foundations via Convolutional Neural Networks. Sensors 2020, 20, 3429. [Google Scholar] [CrossRef]
  28. Zhao, H.; Chen, Z.; Jiang, H.; Jing, W.; Sun, L.; Feng, M. Evaluation of three deep learning models for early crop classification using sentinel-1A imagery time series—A case study in Zhanjiang, China. Remote Sens. 2019, 11, 2673. [Google Scholar] [CrossRef] [Green Version]
  29. Chung, J.; Gulcehre, C.; Cho, K.; Bengio, Y. Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv 2014, arXiv:1412.3555. [Google Scholar]
  30. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  31. Abadi, M.; Agarwal, A.; Barham, P.; Brevdo, E.; Chen, Z.; Citro, C.; Corrado, G.S.; Davis, A.; Dean, J.; Devin, M.; et al. TensorFlow: Large-Scale Machine Learning on Heterogeneous Systems. 2015. Available online: tensorflow.org (accessed on 15 July 2021).
  32. Keras. 2015. Available online: https://keras.io (accessed on 15 July 2021).
  33. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  34. Zhou, P.; Yuan, M.; Wang, H.; Chai, T. Data-driven dynamic modeling for prediction of molten iron silicon content using ELM with self-feedback. Math. Probl. Eng. 2015, 2015, 326160. [Google Scholar] [CrossRef] [Green Version]
  35. Zhong, L.; Hu, L.; Zhou, H. Deep learning based multi-temporal crop classification. Remote Sens. Environ. 2019, 221, 430–443. [Google Scholar] [CrossRef]
  36. Schnaubelt, M. A Comparison of Machine Learning Model Validation Schemes for Non-Stationary Time Series Data; Technical Report, FAU Discussion Papers in Economics, No. 11/2019, Friedrich-Alexander-Universität Erlangen-Nürnberg; Institute for Economics: Nürnberg, Germany, 2019. [Google Scholar]
Figure 1. Steps in the Ferronickel extraction process.
Figure 1. Steps in the Ferronickel extraction process.
Sensors 21 06894 g001
Figure 2. Panoramic of the furnace.
Figure 2. Panoramic of the furnace.
Sensors 21 06894 g002
Figure 3. LSTM internal unit configuration.
Figure 3. LSTM internal unit configuration.
Sensors 21 06894 g003
Figure 4. GRU internal unit configuration.
Figure 4. GRU internal unit configuration.
Sensors 21 06894 g004
Figure 5. General steps in the temperature prediction methodology.
Figure 5. General steps in the temperature prediction methodology.
Sensors 21 06894 g005
Figure 6. External view of panel #17 of the 72 panels wallside that radially compose the furnace.
Figure 6. External view of panel #17 of the 72 panels wallside that radially compose the furnace.
Sensors 21 06894 g006
Figure 7. (Left) radial location of the 4 panels to evaluate and (Right) detail side view of each panel with its 4 levels of plate coolers. Sixteen thermocouples in total were used as output variables.
Figure 7. (Left) radial location of the 4 panels to evaluate and (Right) detail side view of each panel with its 4 levels of plate coolers. Sixteen thermocouples in total were used as output variables.
Sensors 21 06894 g007
Figure 8. Data cleansing process workflow.
Figure 8. Data cleansing process workflow.
Sensors 21 06894 g008
Figure 9. Data cleansing process rules: (a) variables with high variance and (b) strings.
Figure 9. Data cleansing process rules: (a) variables with high variance and (b) strings.
Sensors 21 06894 g009
Figure 10. Data cleansing process rules: (a) remove duplicates and (b) empty and null values.
Figure 10. Data cleansing process rules: (a) remove duplicates and (b) empty and null values.
Sensors 21 06894 g010
Figure 11. Data cleansing process rules: (a) variables which remain in same value for long time periods, (b) unique values, and (c) negative temperature cases.
Figure 11. Data cleansing process rules: (a) variables which remain in same value for long time periods, (b) unique values, and (c) negative temperature cases.
Sensors 21 06894 g011
Figure 12. Output shape in the GRU + Dense model.
Figure 12. Output shape in the GRU + Dense model.
Sensors 21 06894 g012
Figure 13. Steps for the development of the temperature prediction deep learning models.
Figure 13. Steps for the development of the temperature prediction deep learning models.
Sensors 21 06894 g013
Figure 14. True versus predictive behavior of the gated recurrent unit (GRU) model. Training and test sets in one of the output thermocouples.
Figure 14. True versus predictive behavior of the gated recurrent unit (GRU) model. Training and test sets in one of the output thermocouples.
Sensors 21 06894 g014
Figure 15. True and predicted behavior in each one of the 16 thermocouples studied in the test set.
Figure 15. True and predicted behavior in each one of the 16 thermocouples studied in the test set.
Sensors 21 06894 g015
Figure 16. Changing of the loss value through the epocs in train and test sets of the GRU (250 Cells) + Dense model.
Figure 16. Changing of the loss value through the epocs in train and test sets of the GRU (250 Cells) + Dense model.
Sensors 21 06894 g016
Figure 17. Four-fold moving origin cross validation strategy in the time series approach.
Figure 17. Four-fold moving origin cross validation strategy in the time series approach.
Sensors 21 06894 g017
Figure 18. Average RMSE of the train set varying the learning variables at the input of the model.
Figure 18. Average RMSE of the train set varying the learning variables at the input of the model.
Sensors 21 06894 g018
Figure 19. Average RMSE of the test set varying the learning variables at the input of the model.
Figure 19. Average RMSE of the test set varying the learning variables at the input of the model.
Sensors 21 06894 g019
Table 1. Description of input variables.
Table 1. Description of input variables.
Variable IdentificationNumber of Input Variables
calcine feed totalizer1
automatic furnace regulation system1
electrode current3
electrode voltage3
relative electrode position3
electrode arc3
electric furnace power1
electrode power3
chemical composition of calcine15
16 thermocouples to predict16
Total input variables49
Table 2. Amount of variables eliminated in each of the categories after performing the data cleaning process.
Table 2. Amount of variables eliminated in each of the categories after performing the data cleaning process.
CategoryDeleted Variables
High variance97
Strings5
Remove duplicates80
Empty and null values2
Zero variance22
Unique values60
Negative temperatures74
Total deleted variables340
Table 3. Parameters setup in the GRU + Dense model.
Table 3. Parameters setup in the GRU + Dense model.
ParameterValue
Neural network layer typeGRU + Dense
Number of GRU grid cells250
Optimization methodadam
Learning rate 1 × 10 3
Number of training epochs50
Steps per epoch100
Total number of data40000
Data division percentage for training90%
Data division percentage for testing10%
Number of random sequences belonging to the batch for training250
Number of steps in each random sequence1152
Number of shift steps8 every 15 min,
predicts 2 h into the future
Number of input variables49
Number of output variables16
Table 4. RMSE comparison of six different deep learning models for each one of the 16 output thermocouples in the training and test sets. (The percentage of data split in train set was 90% and the remaining 10% was used for test set.)
Table 4. RMSE comparison of six different deep learning models for each one of the 16 output thermocouples in the training and test sets. (The percentage of data split in train set was 90% and the remaining 10% was used for test set.)
GRU 250GRU 128 + GRU 64LSTM 250GRU 128 + LSTM 128CONV1D 128CONV1D 128 + LSTM 128
ThermocoupleTrainTestTrainTestTrainTestTrainTestTrainTestTrainTest
Level 1 NW0.661.341.031.860.801.810.951.820.881.640.871.74
Level 1 SW0.871.101.461.831.091.771.271.891.111.241.151.56
Level 1 SE0.650.710.890.850.761.150.871.020.780.730.720.82
Level 1 NE0.450.430.750.600.580.610.640.590.600.530.600.52
Level 2 NW1.081.461.581.801.312.051.472.461.541.991.382.03
Level 2 SW1.091.541.882.411.412.191.622.701.581.681.452.26
Level 2 SE0.730.891.101.320.841.410.981.430.941.180.831.12
Level 2 NE0.500.430.740.590.650.730.710.690.640.540.610.74
Level 3 NW0.810.851.131.050.951.211.031.241.001.070.981.42
Level 3 SW0.800.771.281.161.011.201.171.331.021.041.061.10
Level 3 SE0.760.961.091.210.881.470.981.480.911.010.801.05
Level 3 NE0.540.470.760.620.620.740.690.660.660.700.620.55
Level 4 NW2.402.943.283.832.863.943.034.462.993.532.653.08
Level 4 SW2.282.613.353.492.773.883.103.633.023.112.623.38
Level 4 SE1.671.592.132.091.942.332.012.261.741.851.801.93
Level 4 NE0.941.021.411.711.141.841.361.891.291.861.161.59
Average1.011.191.491.651.231.771.371.851.291.481.211.56
Table 5. Average RMSE behavior in training and test sets in five different iterations for the GRU (250 Cells) + Dense model. The number of sequences in batch train was set to 100 in all cases.
Table 5. Average RMSE behavior in training and test sets in five different iterations for the GRU (250 Cells) + Dense model. The number of sequences in batch train was set to 100 in all cases.
Iteration 1Iteration 2Iteration 3Iteration 4Iteration 5
Train1.071.071.11.041.1
Test1.241.261.261.231.26
Table 6. Average RMSE behavior in train and test sets changing the number of random sequences for training the GRU (250 Cells) + Dense model.
Table 6. Average RMSE behavior in train and test sets changing the number of random sequences for training the GRU (250 Cells) + Dense model.
Number of Random Sequences for Training
1050100150200250
Train1.081.061.11.051.061.01
Test1.261.221.261.31.291.19
Table 7. Average RMSE moving origin four-fold cross validation results in the training and test sets for the GRU (250 Cells) + Dense model.
Table 7. Average RMSE moving origin four-fold cross validation results in the training and test sets for the GRU (250 Cells) + Dense model.
Fold 1Fold 2Fold 3Fold 4
Train0.520.70.951.02
Test4.932.311.641.60
Table 8. Average RMSE sensitiveness analysis of three different training/test ratios 90/10, 85/15 and 80/20.
Table 8. Average RMSE sensitiveness analysis of three different training/test ratios 90/10, 85/15 and 80/20.
90/1085/1580/20
Train1.011.001.02
Test1.191.541.60
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Leon-Medina, J.X.; Camacho, J.; Gutierrez-Osorio, C.; Salomón, J.E.; Rueda, B.; Vargas, W.; Sofrony, J.; Restrepo-Calle, F.; Pedraza, C.; Tibaduiza, D. Temperature Prediction Using Multivariate Time Series Deep Learning in the Lining of an Electric Arc Furnace for Ferronickel Production. Sensors 2021, 21, 6894. https://doi.org/10.3390/s21206894

AMA Style

Leon-Medina JX, Camacho J, Gutierrez-Osorio C, Salomón JE, Rueda B, Vargas W, Sofrony J, Restrepo-Calle F, Pedraza C, Tibaduiza D. Temperature Prediction Using Multivariate Time Series Deep Learning in the Lining of an Electric Arc Furnace for Ferronickel Production. Sensors. 2021; 21(20):6894. https://doi.org/10.3390/s21206894

Chicago/Turabian Style

Leon-Medina, Jersson X., Jaiber Camacho, Camilo Gutierrez-Osorio, Julián Esteban Salomón, Bernardo Rueda, Whilmar Vargas, Jorge Sofrony, Felipe Restrepo-Calle, Cesar Pedraza, and Diego Tibaduiza. 2021. "Temperature Prediction Using Multivariate Time Series Deep Learning in the Lining of an Electric Arc Furnace for Ferronickel Production" Sensors 21, no. 20: 6894. https://doi.org/10.3390/s21206894

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop