Next Article in Journal
Optimal Scheduling of Microgrid Considering the Interruptible Load Shifting Based on Improved Biogeography-Based Optimization Algorithm
Previous Article in Journal
Plant Leaves Recognition Based on a Hierarchical One-Class Learning Scheme with Convolutional Auto-Encoder and Siamese Neural Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Improved Equilibrium Optimizer Algorithm and Its Application in LSTM Neural Network

School of Electronics and Information Engineering, Hebei University of Technology, Tianjin 300401, China
*
Authors to whom correspondence should be addressed.
Symmetry 2021, 13(9), 1706; https://doi.org/10.3390/sym13091706
Submission received: 6 May 2021 / Revised: 5 September 2021 / Accepted: 6 September 2021 / Published: 15 September 2021

Abstract

:
An improved equilibrium optimizer (EO) algorithm is proposed in this paper to address premature and slow convergence. Firstly, a highly stochastic chaotic mechanism is adopted to initialize the population for range expansion. Secondly, the capability to conduct global search to jump out of local optima is enhanced by assigning adaptive weights and setting adaptive convergence factors. In addition 25 classical benchmark functions are used to validate the algorithm. As revealed by the analysis of the accuracy, speed, and stability of convergence, the IEO algorithm proposed in this paper significantly outperforms other meta-heuristic algorithms. In practice, the distribution is asymmetric because most logging data are unlabeled. Traditional classification models have difficulty in accurately predicting the location of oil layer. In this paper, the oil layers related to oil exploration are predicted using long short-term memory (LSTM) networks. Due to the large amount of data used, however, it is difficult to adjust the parameters. For this reason, an improved equilibrium optimizer algorithm (IEO) is applied to optimize the parameters of LSTM for improved performance, while the effective IEO-LSTM is applied for oil layer prediction. As indicated by the results, the proposed model outperforms the current popular optimization algorithms including particle swarm algorithm PSO and genetic algorithm GA in terms of accuracy, absolute error, root mean square error and mean absolute error.

1. Introduction

Most of the optimization problems arising from various engineering settings are highly complex and challenging. Since traditional approaches tend to use specific solutions to solve specific problems, they are highly characteristic but not universally applicable. Characterized by linear programming and convex optimization [1], traditional methods are restricted by various drawbacks. For example, the problem-solving process is extremely complicated, and it is prone to obtaining local optimal solutions, thus it is difficult to solve realistic problems using traditional methods and therefore, currently it is a common practice to use meta-heuristic optimization algorithms. The meta-heuristic optimization algorithm is becoming popular because it is easy to implement, bypass local optimization, and can be used in a wide range of problems covering different disciplines utilized in various problems covering different disciplines. Meta-heuristic algorithms solve optimization problems by mimicking biological or physical phenomena. It can be grouped into three main categories: natural evolution (EA)-based algorithms [2], physics-based algorithms [3], and swarm intelligence algorithms (SI) [4]. They all have excellent performances in combinatorial optimization [5], feature selection [6], image processing [7], data mining [8], and many other fields. In recent years, new meta-heuristic algorithms based on mimic the natural behavior have been proposed one after another, including particle swarm optimization (PSO) [9,10], gray wolf optimization (GWO) [11], seagull optimization algorithm (SOA) [12], whale optimization algorithm (WOA) [13,14], cuckoo search algorithm (CSA) [15,16], marine predator algorithm (MPA) [17,18], coyote optimization algorithm (COA) [19], carnivorous plant algorithm (CPA) [20], transient search algorithm (TSA) [21], genetic algorithms (GA) [22] and more.
Inspired by control volume mass balance models used to estimate both dynamic and equilibrium states, the equilibrium optimization algorithm (EO) is proposed by Faramarzia and Heidarinejad in 2019 [23]. The mass balance equation is used to describe the concentration of non-reactive components in the control volume depending on the exact resource library model, before reaching the equilibrium state as the optimal result. However, the EO algorithm is disadvantaged by its slow convergence and the tendency to fall into the local optimum. Thus, an improved algorithm is proposed in this paper. Firstly, allowing for the instability caused by the traditional randomly initialized population, chaotic mapping is performed to initialize the population. At this time, the distribution of search agents will be made more comprehensive. Secondly, adaptive weights and adaptive convergence factors are introduced to improve global search ability in the early stage, for avoiding the prospect of falling into the local optimum. In addition, the local search ability in the later stage is also improved. At present, there have been a lot of comparative experiments conducted to validate the improved algorithm. Herein, there are three popular group optimization algorithms chosen for comparison. Besides, there are fifteen commonly-used standard test functions selected, including five high-dimensional unimodal functions, three multimodal functions, seven fixed-dimensional multimodal functions, and 10 composite functions. For each function, 30 experiments were performed, and then the experimental results were recorded. According to the experimental results obtained, IEO has a higher convergence speed and higher search accuracy for most functions.
Like other neural network models, in LSTM applications it is necessary to manually set some parameters for the LSTM neural network model, such as the learning rate and the number of batches. These parameters tend to have a significant impact on the topology of the network model. Since prediction performance of models trained with different parameters varies significantly. To confirm the appropriate model parameters, this paper proposes the EO algorithm with adaptive weights based on chaos theory to optimize the internal parameters of LSTM, and then use the optimized neural network for oil layer prediction. Experimentally, it is proved that the improved EO algorithm can globally optimize parameters of LSTM, reduce the randomness and improve the prediction effect of LSTM for the oil layer.

2. Equilibrium Optimizer Algorithm and Its Improvement

2.1. Description of Equilibrium Optimizer Algorithm

In the equilibrium optimization algorithm, the solution is similar to particles, the concentration is similar to the position of particles in PSO. The mass balance equation is used to describe the concentration of non-reactive components in the control volume. The mass balance equation is reflected in the physical process of controlling the mass entering, leaving and generating in the volume. Usually described by a first-order differential equation, its expressed as follows:
V d C d t = Q C eq Q C + G
where V represents control volume; V d C d t denotes the rate of quality change; C indicates the concentration in the control volume; Q is referred to as the volumetric flow rate into or out of the control volume; Ceq denotes the concentration inside the control volume in the absence of mass generation (which is the equilibrium state); G represents the rate of mass production inside the control volume.
Through the differential equation as described using Equation (1) is solved, it can be obtained that:
C = C eq + C 0 C eq F + G λ V ( 1 F )
F = exp λ t t 0
where C 0 represents the initial concentration of the control volume in time t0; F indicates the index term coefficient; λ refers to the flow rate. For an optimization problem, the concentration C on the left side of the equation represents the newly generated current solution; C 0 denotes the initial concentration of the control volume at the time t 0 , while representing the solution as obtained in the previous iteration; C eq represents the best solution currently found by the algorithm; t is defined as an iterative (iter) function, which decreases as the number of times of iteration increases:
t = 1 ( I t e r M a x _ i t e r ) a 2 I t e r M a x _ i t e r
I t e r and M a x _ i t e r are constant, representing the current and maximum number of times of iteration, respectively; a 2 is a constant, which is used to manage the optimization ability of the function.
The mathematical model is as follows:
(1)
Initialization: In the initial state of the optimization process, the equilibrium concentration is unknown. Therefore, the algorithm performs random initialization within the upper and lower limits of each optimization variable.
C i 0 = C min + r i C max C min , i = 1 , 2 , , n
where C min and C max represent the lower limit and upper limit vector of the optimized variable, respectively; r i denotes the random number vector of individual i, the dimension of which is consistent with that of the optimized space. The value of each element is a random number ranging from 0 to 1.
(2)
Update of the balance pool: The equilibrium state in Equation (2) is selected from the five current optimal candidate solutions to improve the global search capability of the algorithm and avoid the prospect of falling into low-quality local optimal solutions. The equilibrium state pool formed by these candidate solutions is expressed as follows:
C e q , p o a l = C e q ( 1 ) , C e q ( 2 ) , C e q ( 3 ) , C e q ( 4 ) , C e q ( a x e )
where C e q ( 1 ) , C e q ( 2 ) , C e q ( 3 ) , C e q ( 4 ) represent the best solution found through the iteration so far; C e q ( a x e ) refers to the average of these four solutions. The probability is identical for these five candidate solutions selected, that is, 0.2.
(3)
Index term F: To better balance the local search and global search of the algorithm, Equation (3) is improved as follows:
F = a 1 s i g n ( r 0.5 ) e λ t 1
where, a 1 represents the weight constant coefficient of the global search;   s i g n indicates the symbolic function; r , λ both represent a random number vector, the dimension of which is consistent with the optimized space dimension. The value of each element is a random number ranging from 0 to 1.
(4)
Generation rate G: To enhance the local optimization capability of the algorithm, the generation rate is designed as follows:
G = G 0 e λ ( t t 0 ) = G 0 F
where:
G 0 = G C P ( C e q λ C )
G C P = 0.5 r 1 r 2 G P 0 r 2 < G P
when   G P = 0.5 , the algorithm achieves balance between global optimization and local optimization.
(5)
Solution update: To address the update optimization problem, the solution of the algorithm is updated as follows according to Equation (2):
C = C e q + C C e q F + G λ V ( 1 F )
F is obtained using Equation (9), and V represents the unit. The right side of Equation (10) is comprised of three terms. The first one is an equilibrium concentration, while the second and third ones represent the changes in concentration.
Though the equalization optimization algorithm (EO) performs well in optimization, it is prone to find local optimal solutions, as a result of which the capability of global search is inadequate, as is that of convergence [24]. Therefore, an inertial weight optimization algorithm based on chaos theory is proposed in this paper to improve the capability of global search and local convergence.

2.2. Improved Equilibrium Optimizer (IEO)

2.2.1. Improved Strategies for EO

(1)
Population initialization based on chaos theory
The diversity of the initial population can have a considerable impact on the speed and accuracy of convergence for the swarm intelligence algorithm. However, the basic equilibrium optimization algorithm can’t ensure the diversity of the population by randomly initializing the population. Featuring regularity, randomness, and ergodicity, chaotic mapping has been widely practiced for the optimization of intelligent algorithms [25]. In this paper, the information in the solution space is fully extracted and captured through chaotic mapping to enhance the diversity of the initial concentration. One of the mapping mechanisms widely used in the research of chaos theory is chaotic mapping, the mathematical iterative equation of which is expressed as follows:
λ t + 1 = μ × λ t 1 λ t , t = 0 , 1 , 2 , , T
where λ t represents a uniformly distributed random number in the interval [0,1], λ 0 { 0 , 0.25 , 0.5 , 0.75 , 1 } ; T indicates the preset maximum number of chaotic iterations, T D ; μ denotes the chaos control parameter. When μ = 4, the system is in a completely chaotic state.
Equation (8) is first used to generate a set of chaotic variables. Then, the chaotic sequence is adopted to map the concentration of all dimensional liquids into the chaotic interval f min , f max according to Equation (9). Furthermore, the fitness function is used to calculate the fitness value corresponding to each solubility:
X i j = f min + λ j × f max f min
where X i j represents the coordinate of the j-th dimension of the i-th search agent, and λj refers to the coordinate of the j-th dimension randomly sorted inside λ.
(2)
Adaptive inertia weights Based on the nonlinear decreasing strategy
The inertia weight plays an essential role in optimizing the objective function. More specifically, when the inertia weight is larger, the algorithm performs better in global search, and it can search a larger area. Conversely, when the inertia weight is smaller, the algorithm has a better capability of local search, and it can search the area around the optimal solution. However, the value of inertia weight is constant at 1 in most cases of EO initialization. This is not a better value in the early and late stages of the algorithm. Therefore, appropriate weighting is required to improve the capability of optimization for the algorithm. In the process of EO optimization, the convergence speed of the algorithm will be reduced if the linear inertia weight adjustment strategy chosen is inappropriate. Therefore, an approach to adaptively changing the weight based on the number of iterations is proposed in this paper in line with the nonlinear decrement strategy. It is expressed as follows:
w ( t ) = 2 × sin ( ( π × t / ( m × Max _ t ) ) + π ) + n
where m and n represent optional parameters; M a x _ t indicates the maximum number of iterations; t denotes the current number of iterations.
By combining directions Equations (7) and (8), the concentration update formula is expressed as follows:
C = w ( t ) C e q + C C e q F + G λ V ( 1 F )
The fluctuation range of the convergence factor is shown in Figure 1. Plenty of experiments have been conducted to demonstrate that the effect is better when both m and n are 2, respectively.
The large inertia weight and rapid decline in the early stage are conducive to improving the capability of global search. While the small inertia weight and slow decline in the later stage are favorable to enhancing the capability of local search and improving convergence speed. In the later stage of the algorithm, the updated concentration increasingly approaches the optimal concentration to reach an equilibrium state. Thus, the small adaptive weight can be used to change the concentration at this time while improving the capability of local optimization.

2.2.2. IEO Algorithm Process

The main process of the IEO algorithm is detailed as follows:
(1)
The parameters are initialized, including a 1 , a 2 , and G P , Then the maximum number of iterations and the fitness function of the candidate concentration are set;
(2)
The fitness value of each object is calculated according to Equation (5), including: C e q 1 , C e q 2 , C e q 3 , C e q 4 , and C e q ( a v e ) ;
(3)
Calculation is performed for C e q , p o o l according to Equation (6), and the concentration of each search object is updated using Equation (15);
(4)
The output result is saved when the maximum number of iterations is reached.
The pseudo-code of the IEO algorithm is shown in Algorithm 1.
Algorithm 1 Pseudo-code of the IEO algorithm.
  Initialisation:
particle populations i = 1, …., n;
free parameters a1;a2; GP:
a1 = 2;
a2 = 1;
GP = 0.5;
While (Iter < Max_iter) do
    for (i = 1:number of particles(n)) do
    Calculate fitness of i-th particle
          if (fit( C i ) < fit( C e q 1 )) then
    Replace C e q 1 with C i and fit( C e q 1 ) with ( C i )
          Else if (fit( C i ) > fit( C e q 1 ) & fit( C i ) < fit( C e q 2 )) then
             Replace C e q 2 with C i and fit( C e q 2 ) with ( C i );
          Else if (fit( C i ) > fit( C e q 1 ) & fit( C i ) > fit( C e q 2 ) & fit( C i ) < fit( C e q 3 )) then
             Replace C e q 3 with C i and fit( C e q 3 ) with ( C i )
          else fit( C i ) > fit( C e q 1 ) & fit( C i ) > fit( C e q 2 ) & fit( C i ) > fit( C e q 3 ) & fit( C i ) < fit( C e q 4 )
             Replace C e q 4 with C i and fit( C e q 4 ) with ( C i );
      end if
    end for
C a v e = ( C e q 1 + C e q 2 + C e q 3 + C e q 4 ) / 4
Construct the equilibrium pool C e q , p o o l = { C e q 1 , C e q 2 , C e q 3 , C e q 4 , C e q ( a v e ) }
Accomplish memory saving (if Iter > 1)
    Assign t = 1 ( I t e r M a x _ i t e r ) a 2 I t e r M a x _ i t e r
          For I = 1:number of particles(n)
             Randomly choose one candidate from the equilibrium pool(vector)
             Generate random vectors of λ, r
             Construct F = a 1 s i g n ( r 0.5 ) [ e λ t 1 ]
             Construct  G C P = 0.5 r 1 r 2 0.5 0 r 2 < 0.5
             Construct  G 0 = G C P ( C e q λ C )
             Construct  G = G 0 · F
             Update concentrations C = w ( t ) · C e q + ( C C e q ) · F + G λ V ( 1 F )
      end for
    Iter = Iter+1
End while

2.3. Numerical Optimization Experiments

In order to validate the IEO algorithm, it is compared with 6 well-known meta-heuristic algorithms, including equilibrium optimizer (EO), whale optimization algorithm (WOA), marine predator algorithm (MPA), grey wolf optimization algorithm (GWO), particle swarm optimization (PSO), and genetic algorithm (GA). Furthermore, a total of 25 standard test functions are used to test its performance. In the following part, the benchmark function, experimental settings, and simulation results will be detailed.

2.3.1. Experimental Settings and Algorithm Parameters

The experimental environment is Microsoft Corporation, Windows 10 64 bit, MATLAB R2012a, Intel Core CPU (I5-3210M 2.50GHz) equipped with 8 G RAM. Since the results of each algorithm are randomized during operation, each comparison algorithm will be operated independently in all test functions 30 times to ensure the objectivity of comparison. The population size is 30, and the maximum number of iterations is set to 1000. Then, the average and standard deviations are calculated for the data collected.

2.3.2. Benchmark Test Functions

Currently, the parameter settings of these 25 standard test functions have been widely adopted to verify meta-heuristics. In general, it is difficult for an algorithm to fit all test functions. Therefore, the 25 test functions selected are different. In this way, the experimental results can better reflect the capability of optimization for the algorithm.
The five high-dimensional unimodal test functions ( F 1 ( x ) ~ F 5 ( x ) ) have as few as one global optimal value but no local optimal value. Therefore, this function can be applied to test the algorithm for its capability of local search and convergence speed. In contrast to the high-dimensional unimodal function, the three high-dimensional multimodal test functions ( F 6 ( x ) ~ F 8 ( x ) ) have multiple local optimal values, thus making it more difficult for the algorithm to solve high-dimensional multimodal test functions than to solve unimodal functions. Thus, this function can be used to verify the algorithm for its capability of global search. The 7 fixed-dimensional multimodal functions ( F 9 ( x ) ~ F 15 ( x ) ) have multiple local extrema but the level of their dimensionality is lower as compared to high-dimensional multimodal functions. Therefore, the number of local extrema is relatively small. Similar to the high-dimensional multi-modal test function, the fixed-dimensional multi-modal test function can also be used to assess the algorithm for its performance in global search. To test its comprehensive capability, we added 10 composition functions of cec2017 [26]. In Table 1, Dim indicates the dimension of the standard test function, fmin represents the theoretical optimal value of the standard test function, and the range means the range of search space, N means the number of basic functions.

2.3.3. Experimental Results and Analysis

In this section, comparative numerical optimization experiments are conducted on EO, WOA, MPA, GWO, PSO, GA, and IEO, respectively. The convergence curve is presented in Figure A1, and the stability curve of each group optimization algorithm is presented in Figure A2. The test results of the unimodal function, high-dimensional multimodal function, fixed-dimensional multimodal function and composition functions are listed in Table A1, where Ave and Std represent the average solution in 30 independent experiments and the standard deviation of the results in 30 runs, respectively.
As shown in Figure A1, the convergence speed of IEO on F 1 ( x ) ~ F 5 ( x ) is significantly higher when compared with other algorithms and it achieves higher accuracy of convergence. Besides, the convergence of each function is relatively slow, and the IEO convergence value is shown to be the best, suggesting that the IEO is less prone to the local optimum. In general, when the high-dimensional unimodal function is tested, the IEO algorithm demonstrates its superior and consistent performance, implying its effectiveness in solving the high dimensionality of space. On F 6 ( x ) ~ F 8 ( x ) , the convergence speed of IEO is the highest, and the accuracy of convergence is improved, suggesting that IEO has a better ability to jump out of the local optimum than other algorithms. Compared with other algorithms, it can be seen from the average (Ave) and standard deviation (Std) that IEO performs best and is most stable. Overall, the performance of the IEO algorithm in solving the high-dimensional multi-peak test function shows higher stability than other algorithms, which evidences its strong capability of global search. On F 9 ( x ) ~ F 15 ( x ) , in the first five functions, IEO is not only reflected in the fastest convergence speed, but also the higher convergence accuracy. While on F 15 ( x ) , despite the slow convergence of IEO, it jumped out of the local optimum in the later stage and found the optimum value which is closer to the theory. In general, the capability of IEO to jump out of the local optimal value is better than that of other algorithms for the fixed-dimensional multi-modal function. Thus, the algorithm is verified as effective and stable. On F 16 ( x ) ~ F 25 ( x ) , Overall, the convergence speed of IEO is faster than that of EO, and compared with other algorithms, the convergence speed of IEO has certain advantages. Among them, F 16 ( x ) , F 18 ( x ) , F 21 ( x ) and F 23 ( x ) converge the fastest and obtain the optimal value. As shown in Table A1, on F 1 ( x ) ~ F 5 ( x ) , the IEO algorithm outperforms other algorithms. In these five test functions, IEO produces better results than other algorithms regardless of the average value or the standard deviation, and the accuracy of convergence is significantly improved relative to other algorithms. Moreover, in the tests in four of the functions, IEO achieved the ideal optimal value of 0 every time. The variance of the IEO algorithm is considerably smaller as compared to other algorithms, which indicates the stability of the IEO algorithm. On the high-dimensional multimodal test functions F 6 ( x ) ~ F 8 ( x ) , IEO outperforms other algorithms to a significant extent. Compared with other algorithms, IEO performs more consistently with the smallest average and standard deviation. Notably, it achieves the theoretical optimal value of 0 on F 6 ( x ) and F 8 ( x ) . On fixed-dimensional multimodal functions ( F 9 ( x ) ~ F 15 ( x ) ), the data comparison of all algorithms to optimize the fixed-dimensional multimodal function is performed. According to the comparison of the mean and standard deviation in the table, IEO performs better in mean and standard deviation on the test functions of F 9 ( x ) , F 12 ( x ) , and F 13 ( x ) , while MPA performs best in average and standard deviation on F 10 ( x ) , followed by IEO. On F 11 ( x ) , the average and standard deviation of IEO are not as satisfactory as the four functions of MPA, GWO, PSO and GA. On F 14 ( x ) , EO has achieved the best average and standard deviation, followed by IEO. On F 15 ( x ) , IEO and MPA have achieved the best average and standard deviation together. As a whole, IEO shows stronger capabilities of global search in fixed-dimensional multi-modal functions. On F 16 ( x ) ~ F 25 ( x ) , F 16 ( x ) , F 18 ( x ) , F 21 ( x ) and F 22 ( x ) all achieved the minimum value. In the remaining functions, IEO performed better than EO, indicating that the IEO improvement strategy was successful. It shows that IEO has a significant improvement over EO in solving composition functions, and has a great advantage over other algorithms, and ranks second in most of the functions. Figure A2 shows the test result block diagram of the selected function. In the experiment, box plots are used to indicate the distribution of the results for each function after 30 independent tests to demonstrate the stability of IEO and other algorithms in a more intuitive way. The numerical distribution of the selected function is selected from the test functions. IEO exhibits outstanding stability in all functions, which allows it to outperform other algorithms. More specifically, its bottom edge, top edge, and median value of the same function are superior or inferior to other algorithms.
In general, the IEO optimization algorithm outperforms other algorithms in optimization speed, accuracy and stability.

3. LSTM Model Optimization by Improved IEO Algorithm

Time series data are everywhere, and prediction is an eternal topic for human beings. Gooijer and Hyndman highlight researched published in the International Institute of Forecasters (IIF) and key publications in other journals [27]. From the early traditional autoregressive integrated moving average (ARIMA) linear model to various nonlinear models (e.g., SETAR model). Due to increased computational power and electronic storage capacity, nonlinear forecasting models have grown in a spectacular way and received a great deal of attention during the last decades. The neural networks (NN) model is a prominent example of such nonlinear and nonparametric models that do not make assumptions about the parametric form of the functional relationship between the variables. Several studies have been done relating to the application of NN, such as Caraka and Chen used vector autoregressive (VAR) to improve NN for space-time pollution date forecasting modeling [28], then they used VAR-NN-PSO model to predicting PM2.5 in air [29]. Suhartono and Prastyo applied Generalized Space-Time Autoregressive (GSTAVR) to FFNN for forecasting oil production [30]. In recent years, types of recurrent neural networks (RNN) have been frequently employed for forecasting tasks and have shown promising reslts. This is mainly due to RNNs having the ability to persist information about previous time steps and being able to use that information when processing the current time step. Due to vanishing gradients, RNN is unable to persist long-term dependencies. A long short term memory (LSTM) network is a type of RNN that was introduced to persist long term dependencies [31]. By comparing LSTM with a random forest, a standard deep net, and a simple logistic regression in the large-scale financial market prediction task. It is proved that LSTM performed better than others and was a method naturally suitable for this field [32]. By comparing LSTM with the Prophet model proposed by Facebook, the result showed LSTM performs better on minimum air temperature [33]. ChihHung and Yu-Feng proposed a new forecasting framework with the LSTM model to forecasting bitcoin daily prices. The results revealed its possible applicability in various cryptocurrencies prediction [34]. Helmini and Jihan adopted a special variant LSTM with peephole connections for the sales forecasting tasks, proved that the initial LSTM and improved LSTM both outperform two machine learning models (extreme gradient boosting (XGB) and random forest regressor (RFR)) [35]. To enhance the performance of the LSTM model, Gers and Schmidhuber proposed a novel adaptive “forget gate” that enables an LSTM cell to learn to reset itself at appropriate times [36], Karim and Majumdar added the attention model to detect regions of the input sequence that contribute to the class label through the context vector [37]. Saeed and Li proposed a hybrid model called BLSTM, constructing quality intervals based on the high-quality principle, presented excellent prediction performance [38]. Venskus and Treigys described two methods—the LSTM prediction region learning and the wild bootstrapping for estimation of prediction regions—and the expected results have been achieved when it is used for abnormal marine traffic detection [39]. In order to strengthen the early diagnosis of cardiovascular disease, Zheng and Chen presented an arrhythmia classification method that combines 2-D CNN and LSTM and uses ECG images as the input data for the model [40]. Currently, the most mainstream method for time series forecasting would be LSTM and its variants.

3.1. Basic Model of LSTM

The LSTM model introduces a gating unit into its network topology for gaining control on the degree of impact of current information on the previous information, thus endowing the model with a longer “memory function”. It is made suitable for solving long-term nonlinear sequence forecasting problems. The LSTM network structure consists of input gates, forget gates, output gates, and unit states. The basic structure of the network is illustrated in Figure 2.
LSTM is a purpose-built recurrent neural network. Not only does the design of the “gate” structure avoid the gradient disappearance and gradient explosion caused by traditional recurrent neural networks, it is also effective in learning long-term dependency. Therefore, the LSTM model with memory function shows a clear advantage in dealing with the problems arising from the prediction and classification of time series.
Let the input sequence be x 1 , x 2 , , x T , and the hidden layer state be h 1 , h 2 , , h T . Then, at time t:
f t = σ W f h t 1 + U f x t + b f
i t = σ W i h t 1 + U i x t + b i
g t = t a n h W C h t 1 + U C x t + b C
C t = f t C t 1 + i t g t
o t = σ W o h t 1 + U o x t + b o
h t = o t t a n h C t
where, x t represents the input vector of the LSTM cell; h indicates the cell output vector; f t , i t , and   o t refer to the forget gate, input gate, output gate, respectively; C t denotes the status of the cell unit; t stands for the time; σ and t a n h represent the activation functions; W and b denote the weight and deviation matrix, respectively. U stands for the input weight matrix.
The key to LSTM lies in the cell state C t . It maintains the memory of the unit state at time t , and adjusts it through the forget gate f t and input gate i t . The purpose of the forget gate is to make the cell remember or forget its previous state C t 1 . The input gate is purposed to allow incoming signals to update the unit state or prevent it. The output gate aims to control the output of the unit state C t and then transmit it to the next cell. The internal structure of the LSTM unit is comprised of multiple perceptrons. In general, backpropagation algorithms are the most commonly used training method.
The LSTM neural network framework applied in the design of this paper involves one input layer, three hidden layers (64 neurons in the first layer, 16 neurons in the second layer, and 50 neurons in the third fully connected layer), and one output layer (two neuron output). Using the Adam optimizer, the learning rate is 0.01, the loop is executed 30 times, and the batch size is 256.

3.2. LSTM Based on Improved EO

As mentioned above, the parameters of the LSTM-based model are set manually, which however affects the outcome of the prediction made using the model. Thus, an improved EO-LSTM model is proposed in this paper. The main purpose of doing so is to optimize the relevant parameters of the LSTM through the excellent parameter optimization capabilities of the above-mentioned IEO algorithm for the improved outcome of prediction for the LSTM. In order to ensure the objectivity of the experiment, the LSTM model structure is fixed (the number of nodes in each layer does not modified), The modeling process of the IEO-LSTM model is detailed as follows.
(1)
The relevant parameters are initialized. The parameters of the IEO algorithm are initialized, including: population size, fitness function, are free parameter assignment. The parameters of the LSTM algorithm are initialized by setting the time window, the initial learning rate, the number of initial iterations, and the size of the packet. In this paper, the error is treated as a fitness function, which is expressed as follows:
f i t n e s s = 1 acc ( f ; D ) = 1 1 m i = 1 m I f x i = y i
where D represents the training set, m denotes the number of samples in the training set, f x i indicates the predicted sample label, y i means the original sample label, and I ( f x i = y i ) is a function. If f x i = y i , I ( f x i = y i ) = 1; otherwise, I ( f x i = y i ) = 0.
(2)
The LSTM parameters are set as required to form the corresponding particles. The particle structure is ( a l p h a , n u m _ e p o c h s , b a t c h _ s i z e ). Among them, a l p h a represents the LSTM learning rate, n u m _ e p o c h s indicates the number of iterations, b a t c h _ s i z e suggests the size of the bag. The particles mentioned above are the objects of IEO optimization.
(3)
The concentration is updated according to Equation (15). According to the newly-obtained concentration, calculate the fitness value and then update the individual optimal concentration and the global optimal concentration of the particles.
(4)
If the number of iterations reaches the maximum number of times of iteration, that is, 30, the LSTM model trained on the optimal particles will output the prediction value; otherwise, return to step (3) for continued iteration.
(5)
The optimal parameters are substituted into the LSTM model, for constructing obtaining the IEO-LSTM model.
The pseudo-code of the IEO-LSTM on oil prediction algorithm is shown in Algorithm 2.
Algorithm 2 Pseudo-code of the IEO-LSTM on oil prediction.
Input:
Training samples set: trainset_input, trainset_output
LSTM initialization parameters
Output:
IEO-LSTM model
1. Initialize the LSTM model:
      assign parameters para = [lr,epoch,batch_size] = [0.01,100,256];
      loss = categorical_crossentropy, optimizer = adam
2. Set IEO equalization parameters and Import trainset to Optimization:
    Parameters:para = [lr,epoch,batch_size]
    Ll para = [0.001,1, 1, 1, 1, 1];        % Set lower limit for merit search
    Ul para = [0.01, 100,256, 100, 100,100];    % Set the upper limit of merit seeking
      IEO_para = [ll,ul,trainset_input, trainset_output, num_epoch = 30];
      Get_fitness_position ((IEO)) = length(find((trainset_output - LSTM(trainset-
            input)~ = 0))/length(trainset_output);
            % Find the best Concentration to minimize the LSTM error rate
return para_best to LSTM model;
save (‘IEO-LSTM.mat’,’LSTM’)

3.3. Actual Application in Oil Layer Prediction

3.3.1. Design of Oil Layer Prediction System

In respect of oil logging, reservoir prediction is regarded as crucial for improving the accuracy of oil layer prediction. Herein, we apply IEO-LSTM to oil logging and verify the effectiveness of this algorithm by using oil data provided by three oil fields.
Figure 3 shows the block diagram of the oil layer prediction system based on IEO-LSTM.
The oil layer prediction mainly includes the following processes:
(1)
Data acquisition and preprocessing
In practice, the logging data is classified into two categories after being collected. On the one hand, labeled data as obtained through core sampling and laboratory analysis is generally used as a training set. On the other hand, the unlabeled logging data is treated as the test set. The data pre-processing mainly includes denoising process, normalization process, etc.
Besides, due to the different dimensions of each attribute and value range, these data need to be normalized in the first place, so that the sample data ranges between [0,1]. The normalized logging data are then trained and tested. The formula for sample normalization is expressed as follows:
x ¯ = x x min x max x min
where, x x min , x max ; x min represents the minimum value of the attributes of the data sample, and x max refers to the maximum value of the attribute of the data sample.
(2)
Selection of sample set and attribute reduction
In the selection of the sample set, the selection of the sample set for training is supposed not only to be complete and comprehensive but also to be closely related to the evaluation of the oil. Moreover, the degree to which the prediction of oil can be made varies due to the different attributes of oil logging. In general, there are dozens of logging condition attributes in logging data but not all of them play a decisive role, which makes it necessary to perform attribute reduction. In this paper, the discretization algorithm based on the inflection point is applied and then the reduction method based on the attribute dependence is applied to reduce the logging attributes [41].
(3)
IEO-LSTM modeling
First of all, the LSTM model is established, the training set information is inputted after attribute reduction, and The IEO algorithm is used to find the optimal combination of parameters for the LSTM model to obtain a high-precision IEO-LSTM prediction model.
(4)
Prediction
The trained IEO-LSTM model is used to perform prediction and output the results. Among them, before the test set is inputted into the IEO-LSTM model, the reduced test set must be obtained according to the sample attribute reduction result for normalization according to the normalization principle of sample training.
The pseudo-code of the oil layer prediction system based on IEO-LSTM is shown in Algorithm 3.
Algorithm 3 Pseudo-code of the oil prediction system based on IEO-LSTM.
Input:
Training samples set: trainset_input, trainset_output, testset
Test samples set: testset
Output:
Parameter optimization value
Prediction results of the test samples set
1. Original data initialization, Divide the trainset and testset
2. Initialize the LSTM model:
    assign parameters para = [lr,epoch,batch_size] = [0.01,100,256];
    loss = categorical_crossentropy, optimizer = adama;
3. Set IEO equalization parameters and Import trainset to Optimization:
4. Set IEO equalization parameters and Import trainset to Optimization:
   Parameters:para = [lr,epoch,batch_size]
    ll =[0.001,1, 1];        % Set lower limit for merit search
    ul = [0.01, 100, 256];      % Set the upper limit of merit seeking
    para = IEO[ll,ul,trainset_input, trainset_output, num_epoch = 30];
    Get_fitness_position ( C ) = length(find((trainset_output—LSTM(trainset-
        input)~ = 0))/length(trainset_output);
        % Find the best Concentration to minimize the LSTM error rate
return para_best to LSTM model;
save (‘IEO-LSTM.mat’,’LSTM’);
5. Import test samples set to IEO-LSTM
load(‘IEO-LSTM.mat’)
predictset = IEO-LSTM(testset)
return predictset

3.3.2. Practical Application

To verify the effectiveness and stability of the proposed IEO-LSTM model, we selected logging data from three actual oil wells for the experiment. And the prediction results are compared and evaluated with those of the LSTM, PSO-LSTM [42], GA-LSTM [43] and EO-LSTM models.
The logging data for the three selected actual wells were recorded as o 1 , o 2 and o 3 , Table 2 shows the reduction results obtained after the attribute reduction process for the three wells. It can be seen from the table that there are various redundant condition attributes in the original data. After attribute reduction is performed through a rough set, those significant attributes can be selected, which is conducive to simplifying the complexity of the LSTM model. The data set division of three wells is shown in Table 3. It can be seen from Table 3 that the oil layer and dry layer distribution range in the training set and test set.
Table 4 lists the value range of each attribute after o 1 attribute reduction. Among them, G R , D T , S P , L L D , L L S , D E N , and K represent natural gamma, acoustic time difference, natural potential, deep lateral resistivity, shallow lateral resistivity, compensation density, and potassium, respectively. Besides, Table 5 shows the value range of each attribute after o 2 attribute reduction. Among them, A C , R T , and R X O denote jet lag, intact formation resistivity, and flush zone resistivity, respectively. Table 6 lists the value range of each attribute after o 3 attribute reduction, where N G and R I represent neutron gamma and intrusion zone resistivity, respectively.
The logging curve is obtained after the normalization of the reduced attributes, as shown in Figure 4. The horizontal axis and the vertical axis indicate the depth and the standardized value, respectively.
In order to conduct comparative experiments better, the LSTM model, PSO-LSTM model, GA-LSTM model, and EO-LSTM model are established for comparison against the IEO-LSTM model. Then, these optimized classification models are applied to oil predict using the test set.
In addition to the prediction accuracy of the test set, the following performance indicators are defined to assess the performance of the prediction model:
MSE = 1 m i = 1 m f x i y i 2
R M S E = 1 m a i = 1 m f x i y i 2
MAE = 1 m i = 1 m f x i y i
where, f x i and y i represent the predicted output value and the expected output value, respectively.
RMSE is applied to measure the deviation between the observed value and the true value. MAE is referred to as the average of absolute errors. If the RMSE and MAE are reduced, the performance of the algorithm model is improved. Therefore, RMSE is treated as a criterion applied to evaluate the accuracy of each algorithm model. MAE is often treated as the actual forecast and forecast error because it can better reflect the real state of the forecast value error.
For a comprehensive experimental comparison, the number of times of training is set to 30, based on which the average value of the performance indicators is calculated for each model.
Table 7 shows the prediction ability of each model on the test set of three wells ( o 1 , o 2 , o 3 ). On the whole, the prediction accuracy of IEO-LSTM is most satisfactory. Compared with other models, it performs better in each index, and the accuracy of it is about 5% higher than the original LSTM model on average.

4. Conclusions

In this paper, an improved equalization optimization algorithm is proposed, which applies chaos theory to initialize the population and relies on adaptive weights to enhance the capability of global and local search. Firstly, a total of 30 independent experiments are conducted on 25 benchmarks using six popular swarm intelligence algorithms. According to the experimental results, IEO achieves a significant improvement of convergence speed and convergence accuracy as compared to traditional EO. Compared with PSO, GA, MPA, WOA, and GWO, IEO performs better in convergence speed, accuracy, and stability.
The poor robustness of LSTM is attributed to the randomized generation of such parameters as the learning rate, the number of iterations, and the packet size. To improve the performance of LSTM, IEO and LSTM are combined, with the aforementioned three parameters globally optimized to improve the performance of LSTM in prediction.
Finally, it is applied to well-logging oil prediction for the validation of IEO-LSTM. The data set includes the logging data collected from three different wells. The prediction accuracy of the proposed model is shown to be significantly higher as compared to LSTM. Compared with PSO-LSTM, GA-LSTM, and EO-LSTM, the performance of IEO-LSTM is better. When the number of times of iteration is 200, IEO-LSTM is superior to other classification models in terms of classification accuracy, root mean square error, and mean absolute error. In comparison with the actual oil layer distribution, the prediction of IEO-LSTM shows an average classification error of about 5% relative to the original oil layer distribution.
Moreover, this method can also be applied to solve the generic optimization problems encountered in other fields, which suggests a promising prospect of practical application for it. However, LSTM is a special form of time recursive neural network, which appears to solve a fatal defect of RNN, and it still has its own defects: It only solves the gradient problem of RNN to some extent, it is still difficult for data with too large magnitude; It is time-consuming and computationally intensive during optimization. These two problems will be addressed and practically applied in my subsequent research.

Author Contributions

Conceptualization, P.L. and K.X.; methodology, P.L. and K.X.; software, P.L.; validation, P.L., Y.P. and S.F.; formal analysis, P.L. and K.X.; investigation, P.L. and K.X.; resources, P.L., Y.P. and S.F.; data curation, P.L., Y.P. and S.F.; writing—original draft preparation, P.L.; visualization, Y.P. and S.F. supervision, K.X. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (No. U1813222, No. 42075129), Hebei Province Natural Science Foundation (No. E2021202179), Key Research and Development Project from Hebei Province (No. 19210404D, No. 20351802D, No. 21351803D).

Data Availability Statement

All data are available upon request from the corresponding author.

Conflicts of Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Appendix A

The experimental results in Section 2 are presented below, including statistics for each algorithm, convergence curves, and box line plots.
Figure A1. Convergence curves for 25 benchmark functions in Table 1. (a) Convergence curves for F1; (b) Convergence curves for F2; (c) Convergence curves for F3; (d) Convergence curves for F4; (e) Convergence curves for F5; (f) Convergence curves for F6; (g) Convergence curves for F7; (h) Convergence curves for F8; (i) Convergence curves for F9; (j) Convergence curves for F10; (k) Convergence curves for F11; (l) Convergence curves for F12; (m) Convergence curves for F13; (n) Convergence curves for F14; (o) Convergence curves for F15; (p) Convergence curves for F16; (q) Convergence curves for F17; (r) Convergence curves for F18; (s) Convergence curves for F19; (t) Convergence curves for F20; (u) Convergence curves for F21; (v) Convergence curves for F22; (w) Convergence curves for F23; (x) Convergence curves for F24; (y) Convergence curves for F25.
Figure A1. Convergence curves for 25 benchmark functions in Table 1. (a) Convergence curves for F1; (b) Convergence curves for F2; (c) Convergence curves for F3; (d) Convergence curves for F4; (e) Convergence curves for F5; (f) Convergence curves for F6; (g) Convergence curves for F7; (h) Convergence curves for F8; (i) Convergence curves for F9; (j) Convergence curves for F10; (k) Convergence curves for F11; (l) Convergence curves for F12; (m) Convergence curves for F13; (n) Convergence curves for F14; (o) Convergence curves for F15; (p) Convergence curves for F16; (q) Convergence curves for F17; (r) Convergence curves for F18; (s) Convergence curves for F19; (t) Convergence curves for F20; (u) Convergence curves for F21; (v) Convergence curves for F22; (w) Convergence curves for F23; (x) Convergence curves for F24; (y) Convergence curves for F25.
Symmetry 13 01706 g0a1aSymmetry 13 01706 g0a1bSymmetry 13 01706 g0a1c
Figure A2. Stability curves of various swarm optimization algorithms,“+” Represents the outliers in the calculation results. (a) Box line diagram for F1; (b) Box line diagram for F2; (c) Box line diagram for F3; (d) Box line diagram for F4; (e) Box line diagram for F5; (f) Box line diagram for F6; (g) Box line diagram for F7; (h) Box line diagram for F8; (i) Box line diagram for F9; (j) Box line diagram for F10; (k) Box line diagram for F11; (l) Box line diagram for F12; (m) Box line diagram for F13; (n) Box line diagram for F14; (o) Box line diagram for F15; (p) Box line diagram for F16; (q) Box line diagram for F17; (r) Box line diagram for F18; (s) Box line diagram for F19; (t) Box line diagram for F20; (u) Box line diagram for F21; (v) Box line diagram for F22; (w) Box line diagram for F23; (x) Box line diagram for F24; (y) Box line diagram for F25.
Figure A2. Stability curves of various swarm optimization algorithms,“+” Represents the outliers in the calculation results. (a) Box line diagram for F1; (b) Box line diagram for F2; (c) Box line diagram for F3; (d) Box line diagram for F4; (e) Box line diagram for F5; (f) Box line diagram for F6; (g) Box line diagram for F7; (h) Box line diagram for F8; (i) Box line diagram for F9; (j) Box line diagram for F10; (k) Box line diagram for F11; (l) Box line diagram for F12; (m) Box line diagram for F13; (n) Box line diagram for F14; (o) Box line diagram for F15; (p) Box line diagram for F16; (q) Box line diagram for F17; (r) Box line diagram for F18; (s) Box line diagram for F19; (t) Box line diagram for F20; (u) Box line diagram for F21; (v) Box line diagram for F22; (w) Box line diagram for F23; (x) Box line diagram for F24; (y) Box line diagram for F25.
Symmetry 13 01706 g0a2aSymmetry 13 01706 g0a2bSymmetry 13 01706 g0a2c
Table A1. Optimization results and comparison for functions.
Table A1. Optimization results and comparison for functions.
Function IEOEOWOAMPAGWOPSOGA
Unimodal Benchmark FunctionsF1Ave0.00 × 10+002.45 × 10−851.41 × 10−303.27 × 10−211.07 × 10−580.32 × 10+000.55 × 10+00
Std0.00 × 10+009.08 × 10−854.91 × 10−304.61 × 10−212.77 × 10−580.21 × 10+001.23 × 10+00
F2Ave0.00 × 10+002.91 × 10−481.06 × 10−210.07 × 10+001.36 × 10−341.04 × 10+000.01 × 10+00
Std0.00 × 10+002.92 × 10−482.39 × 10−210.11 × 10+001.40 × 10−340.46 × 10+000.01 × 10+00
F3Ave0.00 × 10+001.77 × 10−215.39 × 10−072.78 × 10+029.25 × 10−168.14 × 10+018.46 × 10+02
Std0.00 × 10+008.45 × 10−212.93 × 10−064.00 × 10+022.44 × 10−152.13 × 10+011.62 × 10+02
F4Ave0.00 × 10+002.84 × 10−210.73 × 10−016.78 × 10+001.84 × 10−141.51 × 10+004.56 × 10+00
Std0.00 × 10+009.26 × 10−210.40 × 10+002.94 × 10+003.40 × 10−140.22 × 10+000.59 × 10+00
F5Ave6.70 × 10−050.00 × 10+000.00 × 10+000.00 × 10+000.09 × 10+000.07 × 10+000.11 × 10+00
Std0.00 × 10+000.00 × 10+000.00 × 10+000.00 × 10+000.03 × 10+000.01 × 10+000.04 × 10+00
Multimodal Benchmark FunctionsF6Ave0.00 × 10+000.00 × 10+000.00 × 10+000.00 × 10+000.00 × 10+004.84 × 10+010.00 × 10+00
Std0.00 × 10+000.00 × 10+000.00 × 10+000.00 × 10+000.00 × 10+003.44 × 10+003.97 × 10+00
F7Ave8.84 × 10−165.15 × 10−157.40 × 10+009.69 × 10−120.00 × 10+001.20 × 10+000.18 × 10+00
Std0.00 × 10+001.45 × 10−159.90 × 10+006.13 × 10−120.00 × 10+000.73 × 10+000.15 × 10+00
F8Ave0.00 × 10+000.00 × 10+000.00 × 10+000.00 × 10+000.60 × 10+000.01 × 10+000.66 × 10+00
Std0.00 × 10+000.00 × 10+000.00 × 10+000.00 × 10+000.09 × 10+000.01 × 10+000.19 × 10+00
Fifixed-Dimenstion Multimodal Benchmark FunctionsF9Ave1.00 × 10+001.00 × 10+002.11 × 10+001.00 × 10+000.16 × 10+002.18 × 10+001.00 × 10+00
Std2.12 × 10−171.11 × 10−162.50 × 10+002.47 × 10−160.84 × 10+002.01 × 10+008.84 × 10−12
F10Ave3.58 × 10−043.69 × 10−490.00 × 10+003.07 × 10−041.62 × 10−045.61 × 10−042.69 × 10−03
Std1.54 × 10−042.36 × 10−040.00 × 10+004.09 × 10−153.33 × 10−044.38 × 10−044.84 × 10−03
F11Ave−1.03 × 10+00−1.03 × 10+00−1.03 × 10+00−1.03 × 10+00−1.03 × 10+00−1.03 × 10+00−1.03 × 10+00
Std5.63 × 10−062.04 × 10+164.2 × 10−074.46 × 10−015.46 × 10−096.64 × 10−162.39 × 10−08
F12Ave3.00 × 10+003.00 × 10+003.00 × 10+003.00 × 10+003.00 × 10+003.00 × 10+003.00 × 10+00
Std5.95 × 10−174.96 × 10−164.22 × 10−151.95 × 10−152.29 × 10−021.38 × 10−152.45 × 10−07
F13Ave−3.86 × 10+00−3.86 × 10+00−3.85 × 10+00−3.86 × 10+000.51 × 10+00−3.86 × 10+00−3.86 × 10+00
Std3.17 × 10−162.15 × 10−150.00 × 10+002.42 × 10−152.73 × 10−152.68 × 10−152.85 × 10−8
F14Ave−3.32 × 10+00-−3.28 × 10+00−2.98 × 10+00−3.32 × 10+00−2.84×10+00−3.26 × 10+00−3.27 × 10+00
Std1.59 × 10−150.06 × 10+000.38 × 10+001.14 × 10−114.68 × 10−26.05 × 10−25.99 × 10−2
F15Ave−1.05 × 10+1−10.27 × 10+00−9.34 × 10+00−1.05 × 10+1−7.11 × 10+00−7.25 × 10+00−7.77 × 10+00
Std1.92 × 10−071.48 × 10+002.41 × 10+003.89 × 10−113.11 × 10+003.66 × 10+003.73 × 10+00
Composition FunctionsF16Ave2.20 × 10+032.30 × 10+032.33 × 10+032.20 × 10+032.31 × 10+032.32 × 10+032.28 × 10+03
Std6.54 × 10−074.13 × 10+014.34 × 10+018.66 × 10−062.28 × 10+016.34 × 10+015.84 × 10+01
F17Ave2.28 × 10+32.30 × 10+032.36 × 10+032.28 × 10−062.31 × 10+032.37 × 10+032.41 × 10+03
Std0.54 × 10+000.35 × 10+001.43 × 10+023.38 × 10+018.52 × 10+003.20 × 10+021.00 × 10+02
F18Ave2.61 × 10+032.62 × 10+032.65 × 10+032.61 × 10+032.62 × 10+032.71 × 10+032.65 × 10+03
Std3.55 × 10+008.76 × 10+002.88 × 10+015.06 × 10+001.04 × 10+014.18 × 10+011.28 × 10+01
F19Ave2.63 × 10+022.74 × 10+032.76 × 10+032.51 × 10+032.75 × 10+032.77 × 10+022.79 × 10+03
Std5.13 × 10+004.30 × 10+006.11 × 10+013.58 × 10+011.40 × 10+011.16 × 10+022.31 × 10+01
F20Ave2.91 × 10+032.92 × 10+032.95 × 10+032.88 × 10+032.94 × 10+032.90 × 10+032.99 × 10+02
Std2.29 × 10+012.39 × 10+022.36 × 10+012.09 × 10+012.68 × 10+018.60 × 10+018.07 × 10+01
F21Ave2.78 × 10+033.03 × 10+033.60 × 10+032.79 × 10+033.23 × 10+033.08 × 10+033.20 × 10+03
Std5.11 × 10+013.07 × 10+026.33 × 10+021.10 × 10+024.32 × 10+023.57 × 10+022.17 × 10+02
F22Ave3.08 × 10+033.09 × 10+033.13 × 10+033.09 × 10+033.10 × 10+033.16 × 10+033.14 × 10+03
Std0.33 × 10+006.52 × 10+004.27 × 10+010.22 × 10+002.11 × 10+017.44 × 10+013.53 × 10+01
F23Ave3.11 × 10+033.32 × 10+033.41 × 10+033.12 × 10+033.34 × 10+033.19 × 10+033.62 × 10+03
Std5.57 × 10−051.41 × 10+021.77 × 10+025.00 × 10−059.99 × 10+016.07 × 10+012.02 × 10+02
F24Ave3.15 × 10+033.18 × 10+033.36 × 10+033.14 × 10+033.22 × 10+033.26 × 10+033.30 × 10+03
Std3.14 × 10+003.47 × 10+011.12 × 10+028.63 × 10+005.49 × 10+018.77 × 10+018.67 × 10+01
F25Ave4.73 × 10+033.35 × 10+059.75 × 10+053.40 × 10+031.03 × 10+061.08 × 10+041.39 × 10+06
Std3.51 × 10+014.11 × 10+058.97 × 10+053.51 × 10+011.09 × 10+064.96 × 10+031.10 × 10+06

References

  1. Ma, X.S.; Li, Y.L.; Yan, L. A comparative review of traditional multi-objective optimization methods and multi-objective genetic algorithms. Electr. Drive Autom. 2010, 32, 48–53. [Google Scholar]
  2. Koziel, S.; Michalewicz, Z. Evolutionary Algorithms, homomorphous mappings, and constrained parameter optimization. Evol. Comput. 1999, 7, 19–44. [Google Scholar] [CrossRef]
  3. Hashim, F.A.; Houssein, E.H.; Mabrouk, M.S.; Al-Atabany, W.; Mirjalili, S. Henry gas solubility optimization: A novel physics-based algorithm. Future Gener. Comput. Syst. 2019, 101, 646–667. [Google Scholar] [CrossRef]
  4. Karaboga, D.; Basturk, B. A powerful and efficient algorithm for numerical function optimization: Artificial bee colony (ABC) algorithm. J. Glob. Optim. 2007, 39, 459–471. [Google Scholar] [CrossRef]
  5. Fister, I.; Yang, X.S.; Brest, J.; Fister, D. A Brief Review of Nature-Inspired Algorithms for Optimization. Elektrotehniski Vestn. Electrotech. Rev. 2013, 80, 116–122. [Google Scholar]
  6. Brezočnik, L.; Fister, I.; Podgorelec, V. Swarm Intelligence Algorithms for Feature Selection: A Review. Appl. Sci. 2018, 8, 9. [Google Scholar] [CrossRef] [Green Version]
  7. Omran, M.; Engelbrecht, A.; Salman, A. Particle Swarm Optimization Methods for Pattern Recognition and Image Processing. Ph.D. Thesis, University of Pretoria, Pretoria, South Africa, 2004. [Google Scholar]
  8. Martens, D.; Baesens, B.; Fawcett, T. Editorial survey: Swarm intelligence for data mining. Mach. Learn. 2010, 82, 1–42. [Google Scholar] [CrossRef] [Green Version]
  9. Kennedy, J.; Eberhart, R. Particle Swarm Optimization. In Proceedings of the IEEE International Conference on Neural Networks, Perth, Australia, 27 November–1 December 1995; pp. 1942–1948. [Google Scholar]
  10. Zhang, J.; Xia, K.; He, Z.; Fan, S. Dynamic Multi-Swarm Differential Learning Quantum Bird Swarm Algorithm and Its Application in Random Forest Classification Model. Comput. Intell. Neurosci. 2020, 2020, 6858541. [Google Scholar] [CrossRef] [PubMed]
  11. Alejo-Reyes, A.; Cuevas, E.; Rodríguez, A.; Mendoza, A.; Olivares-Benitez, E. An Improved Grey Wolf Optimizer for a Supplier Selection and Order Quantity Allocation Problem. Mathematics 2020, 8, 1457. [Google Scholar] [CrossRef]
  12. Dhiman, G.; Kumar, V. Seagull optimization algorithm: Theory and its applications for large-scale industrial engineering problems. Knowl.-Based Syst. 2018, 165, 169–196. [Google Scholar] [CrossRef]
  13. Mirjalili, S.; Lewis, A. The Whale Optimization Algorithm. Adv. Eng. Softw. 2016, 95, 51–67. [Google Scholar] [CrossRef]
  14. Pan, Y.-K.; Xia, K.-W.; Niu, W.-J.; He, Z.-P. Semisupervised SVM by Hybrid Whale Optimization Algorithm and Its Application in Oil Layer Recognition. Math. Probl. Eng. 2021, 2021, 5289038. [Google Scholar] [CrossRef]
  15. Yang, X.S.; Deb, S. Engineering Optimisation by Cuckoo Search. Int. J. Math. Model. Numer. Optim. 2010, 1, 330–343. [Google Scholar] [CrossRef]
  16. He, Z.; Xia, K.; Niu, W.; Aslam, N.; Hou, J. Semisupervised SVM Based on Cuckoo Search Algorithm and Its Application. Math. Probl. Eng. 2018, 2018, 8243764. [Google Scholar] [CrossRef] [Green Version]
  17. Faramarzi, A.; Heidarinejad, M.; Mirjalili, S.; Gandomi, A.H. Marine Predators Algorithm: A nature-inspired metaheuristic. Expert Syst. Appl. 2020, 152, 113377. [Google Scholar] [CrossRef]
  18. Yang, W.B.; Xia, K.W.; Li, T.J.; Xie, M.; Song, F. A Multi-Strategy Marine Predator Algorithm and Its Application in Joint Regularization Semi-Supervised ELM. Mathematics 2021, 9, 291. [Google Scholar] [CrossRef]
  19. Pierezan, J.; Coelho, L. Coyote Optimization Algorithm: A New Metaheuristic for Global Optimization Problems. In Proceedings of the 2018 IEEE Congress on Evolutionary Computation (CEC), Rio de Janeiro, Brazil, 8–13 July 2018; IEEE: Piscataway, NJ, USA, 2018. [Google Scholar]
  20. Kiong, S.C.; Ong, P.; Sia, C.K. A carnivorous plant algorithm for solving global optimization problems. Appl. Soft Comput. 2020, 98, 106833. [Google Scholar]
  21. Yang, W.B.; Xia, K.W.; Li, T.J.; Xie, M.; Song, F.; Zhao, Y.L. An Improved Transient Search Optimization with Neighborhood Dimensional Learning for Global Optimization Problems. Symmetry 2021, 13, 244. [Google Scholar] [CrossRef]
  22. Qais, M.H.; Hasanien, H.M.; Alghuwainem, S. Transient search optimization: A new meta-heuristic optimization algorithm. Appl. Intell. 2020, 50, 3926–3941. [Google Scholar] [CrossRef]
  23. Afshin, F.; Mohammad, H.; Stephens, B.; Mirjalili, S. Equilibrium optimizer: A novel optimization algorithm. Knowl. Based Syst. 2019, 191, 105190. [Google Scholar]
  24. Yang, L.Q. A novel improved equilibrium global optimization algorithm based on Lévy flight. Astronaut. Meas. Technol. 2020, 40, 66–73. [Google Scholar]
  25. Feng, Y.H.; Liu, J.Q.; He, Y.C. Dynamic population firefly algorithm based on chaos theory. J. Comput. Appl. 2013, 54, 796–799. [Google Scholar]
  26. Wu, G.; Mallipeddi, R.; Ponnuthurai, N. Problem Definitions and Evaluation Criteria for the CEC 2017 Competition on Constrained Real Parameter Optimization; Nanyang Technological University: Singapore, 2010. [Google Scholar]
  27. Gooijer, J.; Hyndman, R.J. 25 years of time series forecasting. In Monash Econometrics & Business Statistics Working Papers; Monash University: Clayton, Australia, 2005; Volume 22, pp. 443–473. [Google Scholar]
  28. Caraka, R.E.; Chen, R.C.; Yasin, H.; Suhartono, S.; Lee, Y.; Pardamean, B. Hybrid Vector Autoregression Feedforward Neural Network with Genetic Algorithm Model for Forecasting Space-Time Pollution Data. Indones. J. Sci. Technol. 2021, 6, 243–268. [Google Scholar] [CrossRef]
  29. Caraka, R.E.; Chen, R.C.; Toharudin, T.; Pardamean, B.; Yasin, H.; Wu, S.H. Prediction of Status Particulate Matter 2.5 using State Markov Chain Stochastic Process and Hybrid VAR-NN-PSO. IEEE Access 2019, 7, 161654–161665. [Google Scholar] [CrossRef]
  30. Suhartono, S.; Prastyo, D.D.; Kuswanto, H.; Lee, M.H. Comparison between VAR, GSTAR, FFNN-VAR and FFNN-GSTAR Models for Forecasting Oil Production. Mat. Malays. J. Ind. Appl. Math. 2018, 34, 103–111. [Google Scholar] [CrossRef]
  31. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  32. Fischer, T.; Krauss, C. Deep learning with long short-term memory networks for financial market predictions. Eur. J. Oper. Res. 2017, 270, 654–669. [Google Scholar] [CrossRef] [Green Version]
  33. Toharudin, T.; Pontoh, R.S.; Caraka, R.E.; Zahroh, S.; Lee, Y.; Chen, R.C. Employing Long Short-Term Memory and Facebook Prophet Model in Air Temperature Forecasting. Commun. Stat. Simul. Comput. 2021. (Early Access). [Google Scholar] [CrossRef]
  34. Wu, C.H.; Lu, C.C.; Ma, Y.-F.; Lu, R.-S. A new forecasting framework for bitcoin price with LSTM. In Proceedings of the IEEE International Conference on Data Mining Workshops, Singapore, 17–20 November 2018; pp. 168–175. [Google Scholar]
  35. Helmini, S.; Jihan, N.; Jayasinghe, M.; Perera, S. Sales Forecasting Using Multivariate Long Short Term Memory Networks. PeerJ 2019, 7, e27712v1. [Google Scholar]
  36. Gers, F.A.; Schmidhuber, J.; Cummins, F. Learning to forget: Continual prediction with LSTM. Neural Comput. 2000, 12, 2451–2471. [Google Scholar] [CrossRef]
  37. Karim, F.; Majumdar, S.; Darabi, H.; Chen, S. LSTM Fully Convolutional Networks for Time Series Classification. IEEE Access 2017, 6, 1662–1669. [Google Scholar] [CrossRef]
  38. Saeed, A.; Li, C.; Danish, M.; Rubaiee, S.; Tang, G.; Gan, Z.; Ahmed, A. Hybrid Bidirectional LSTM Model for Short-Term Wind Speed Interval Prediction. IEEE Access 2020, 8, 182283–182294. [Google Scholar] [CrossRef]
  39. Venskus, J.; Treigys, P.; Markevičiūtė, J. Unsupervised marine vessel trajectory prediction using LSTM network and wild bootstrapping techniques. Nonlinear Anal. Model. Control. 2021, 26, 718–737. [Google Scholar] [CrossRef]
  40. Zheng, Z.; Chen, Z.; Hu, F.; Zhu, J.; Tang, Q.; Liang, Y. An automatic diagnosis of arrhythmias using a combination of CNN and LSTM technology. Electronics 2020, 9, 121. [Google Scholar] [CrossRef] [Green Version]
  41. Bai, J.; Xia, K.; Lin, Y.; Wu, P. Attribute Reduction Based on Consistent Covering Rough Set and Its Application. Complexity 2017, 2017, 8986917. [Google Scholar] [CrossRef] [Green Version]
  42. Wang, P.; Zhao, J.; Gao, Y.; Sotelo, M.A.; Li, Z. Lane Work-schedule of Toll Station Based on Queuing Theory and PSO-LSTM Model. IEEE Access 2020, 8, 84434–84443. [Google Scholar] [CrossRef]
  43. Wen, H.; Zhang, D.; Lu, S.Y. Application of GA-LSTM model in highway traffic flow prediction. J. Harbin Inst. Technol. 2019, 51, 81–95. [Google Scholar]
Figure 1. Fluctuation range of inertia weights.
Figure 1. Fluctuation range of inertia weights.
Symmetry 13 01706 g001
Figure 2. LSTM network structure.
Figure 2. LSTM network structure.
Symmetry 13 01706 g002
Figure 3. Block diagram of the oil layer prediction system based on IEO-LSTM.
Figure 3. Block diagram of the oil layer prediction system based on IEO-LSTM.
Symmetry 13 01706 g003
Figure 4. The normalized curves of attributes: (a,b) attribute normalization of o 1 ; (c,d) attribute normalization of o 2 ; (e) attribute normalization of o 3 .
Figure 4. The normalized curves of attributes: (a,b) attribute normalization of o 1 ; (c,d) attribute normalization of o 2 ; (e) attribute normalization of o 3 .
Symmetry 13 01706 g004
Table 1. Description of benchmark functions.
Table 1. Description of benchmark functions.
FunctionDimRange f min
F 1 ( x ) = i = 1 n x i 2 30[−100, 100]0
F 2 ( x ) = i = 1 n x i + i = 1 n x i 30[−10, 10]0
F 3 ( x ) = i = 1 n j 1 i x j 2 30[−100, 100]0
F 4 ( x ) = max i x i , 1 i n 30[−100, 100]0
F 5 ( x ) = i = 1 n i x i 4 + random [ 0 , 1 ) 30[−1.28, 1.28]0
F 6 ( x ) = i = 1 n x i 2 10 cos 2 π x i + 10 30[−5.12, 5.12]0
F 7 ( x ) = 20 exp 0.2 1 n i = 1 n x i 2 exp 1 n i = 1 n cos 2 π x i + 20 + e 30[−32, 32]0
F 8 ( x ) = 1 4000 i = 1 n x i 2 i = 1 n cos x i i + 1 30[−600, 600]−418.9829 × 5
F 9 ( x ) = 1 500 + j = 1 25 1 j + i = 1 2 x i a i j 6 1 2[−65, 65]1
F 10 ( x ) = i = 1 11 a i x 1 b i 2 + b i x 2 b i 2 + b i x 3 + x 4 2 4[−5, 5]0.00030
F 11 ( x ) = 4 x 1 2 2.1 x 1 4 + 1 3 x 1 6 + x 1 x 2 4 x 2 2 + 4 x 2 4 2[−5, 5]−1.0316
F 12 ( x ) = 1 + x 1 + x 2 + 1 2 19 14 x 1 + 3 x 1 2 14 x 2 + 6 x 1 x 2 + 3 x 2 2 × 30 + 2 x 1 3 x 2 2 × 18 32 x 1 + 12 x 1 2 + 48 x 2 36 x 1 x 2 + 27 x 2 2 2[−2, 2]3
F 13 ( x ) = i = 1 4 c i exp j = 1 3 a i j x j p i j 2 3[1, 3]−3.86
F 14 ( x ) = i = 1 4 c i exp j = 1 6 a i j x j p i j 2 6[0, 1]−3.32
F 15 ( x ) = i = 1 10 X a i X a i T + c i 1 4[0, 10]−10.5363
F 16 ( x ) = C o m p o s i t i o n   F u n c t i o n   1   N = 3 30 100 ,     100 D 0
F 17 ( x ) = C o m p o s i t i o n   F u n c t i o n   2   N = 3 30 100 ,     100 D 0
F 18 ( x ) = C o m p o s i t i o n   F u n c t i o n   3   N = 4 30 100 ,     100 D 0
F 19 ( x ) = C o m p o s i t i o n   F u n c t i o n   4   N = 4 30 100 ,     100 D 0
F 20 ( x ) = C o m p o s i t i o n   F u n c t i o n   5   N = 5 30 100 ,     100 D 0
F 21 ( x ) = C o m p o s i t i o n   F u n c t i o n   6   N = 5 30 100 ,     100 D 0
F 22 ( x ) = C o m p o s i t i o n   F u n c t i o n   7   N = 6 30 100 ,     100 D 0
F 23 ( x ) = C o m p o s i t i o n   F u n c t i o n   8   N = 6 30 100 ,     100 D 0
F 24 ( x ) = C o m p o s i t i o n   F u n c t i o n   9   N = 3 30 100 ,     100 D 0
F 25 ( x ) = C o m p o s i t i o n   F u n c t i o n   10   N = 3 30 100 ,     100 D 0
Table 2. Reduction results of logging data.
Table 2. Reduction results of logging data.
WellAttributes
Original results ( o 1 )GR,DT,SP,WQ,LLD,LLS,DEN,NPHI,PE,U,TH,K,CALI
Reduction results ( o 1 )GR, DT,SP,LLD,LLS, DEN, K
Original results ( o 2 )AC,CNL,DEN,GR,RT,RI,RXO,SP,R2M,R025,BZSP,RA2,C1,C2,CALI,RINC,PORT,VCL,VMA1,VMA6,RHOG,SW,VO,WO,PORE,VXO,VW,so,rnsy,rsly,rny,AC1
Reduction results ( o 2 )AC,GR,RT,RXO,SP
Original results ( o 3 )AC,CALI,GR,NG,RA2,RA4,RI,RM,RT,RXO,SP
Reduction results ( o 3 )AC,NG,RI,SP
Decision attributewhere–1, 1 represents the dry layer and oil layer respectively.
Table 3. Data set partition details.
Table 3. Data set partition details.
WellDepth (m)Training SetDepth (m)Test Set
Oil LayerDry LayerOil LayerDry Layer
o 1 3150~333012213193330~3460115925
o 2 1180~1255654961230~130040360
o 3 1150~1250657361233~1260104112
Table 4. Value range of attributes in o 1 .
Table 4. Value range of attributes in o 1 .
AttributesGRDTSPLLDLLSDENK
Boundary[6, 200][152, 462][−167, −68][0, 25000][0, 3307][1, 4][0, 5]
Table 5. Value range of attributes in o 2 .
Table 5. Value range of attributes in o 2 .
AttributesACGRRTRXOSP
Boundary[54, 140][27, 100][2, 90][1, 328][−32, −6]
Table 6. Value range of attributes in o 3 .
Table 6. Value range of attributes in o 3 .
AttributesACNGRISP
Boundary[0, 571][0, 42][−1, 81][−50, 28]
Table 7. Prediction performance on o 1 , o 2 , o 3 .
Table 7. Prediction performance on o 1 , o 2 , o 3 .
Optimization AlgorithmOilLearning RateEpochsBatch SizeACCMSERMSEMAE
LSTMo10.011002560.88370.20490.48870.2341
o20.011002560.89370.17650.05190.0826
o30.011002560.90950.15610.48960.1971
PSO_LSTMo10.0098861030.90250.16970.39250.1137
o20.0011421540.91540.09370. 0350.0985
o30.00352750.88590.09520.31310.0933
GA_LSTMo10.0082202000.90840.67320.85870.7368
o20.0082541670.89790.07530.28530.0877
o30.0038291760.91320.13680.33680.1404
EO_LSTMo10.0089281630.92310.14810.23470.5442
o20.0081451270.91830.06050.03750.0116
o30.0049561550.92030.07950.23530.0793
IEO_LSTMo10.0030 291370.94330.10230.37810.4544
o20.0082431090.93710.0410.02120.0107
o30.0027531280.96610.03230.20160.0571
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lan, P.; Xia, K.; Pan, Y.; Fan, S. An Improved Equilibrium Optimizer Algorithm and Its Application in LSTM Neural Network. Symmetry 2021, 13, 1706. https://doi.org/10.3390/sym13091706

AMA Style

Lan P, Xia K, Pan Y, Fan S. An Improved Equilibrium Optimizer Algorithm and Its Application in LSTM Neural Network. Symmetry. 2021; 13(9):1706. https://doi.org/10.3390/sym13091706

Chicago/Turabian Style

Lan, Pu, Kewen Xia, Yongke Pan, and Shurui Fan. 2021. "An Improved Equilibrium Optimizer Algorithm and Its Application in LSTM Neural Network" Symmetry 13, no. 9: 1706. https://doi.org/10.3390/sym13091706

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop