Next Article in Journal
A Method for Monthly Extreme Precipitation Forecasting with Physical Explanations
Previous Article in Journal
3D CFD Study of Scour in Combined Wave–Current Flows around Rectangular Piles with Varying Aspect Ratios
Previous Article in Special Issue
Prediction of Sediment Yields Using a Data-Driven Radial M5 Tree Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Inflow Prediction of Centralized Reservoir for the Operation of Pump Station in Urban Drainage Systems Using Improved Multilayer Perceptron Using Existing Optimizers Combined with Metaheuristic Optimization Algorithms

School of Civil Engineering, Chungbuk National University, Cheongju 28644, Republic of Korea
Water 2023, 15(8), 1543; https://doi.org/10.3390/w15081543
Submission received: 14 March 2023 / Revised: 7 April 2023 / Accepted: 12 April 2023 / Published: 14 April 2023

Abstract

:
Owing to the recent increase in abnormal climate, various structural measures including structural and non-structural approaches have been proposed for the prevention of potential water disasters. As a non-structural measure, fast and safe drainage is an essential preemptive operation of a drainage facility, including a centralized reservoir (CRs). To achieve such a preemptive operation, it is necessary to predict the inflow of the drainage facilities. Among the drainage facilities, CRs are located downstream of the drainage area, and their pump stations are operated according to the CR water level. The water level of a CR depends on the inflow, as does the preemptive operation of its pump station. In this study, as a nonstructural measure, the inflow prediction for the CR operation in an urban drainage system was proposed. For predicting the inflow of a CR, a new multilayer perceptron (MLP) using existing optimizers combined with a self-adaptive metaheuristic optimization algorithm, such as an improved harmony search, was proposed. Compared with the adaptive moment, which yields the best results among other existing optimizers, an MLP using an existing optimizer combined with an improved harmony search improves the mean square error and mean absolute error by 0.1767 and 0.0349, respectively.

1. Introduction

To prevent disaster caused by water, structural measures focusing on increasing the capacity of hydraulic facilities, such as pump stations, centralized reservoirs (CRs), decentralized reservoirs (DRs), and pipes, have been proposed. However, owing to the rapidly increasing frequency of localized heavy rainfall, implementable structural measures are limited by time and cost. Hence, non-structural measures, such as the operation of urban drainage systems (UDSs), should be implemented. Among the various non-structural measures available, the operation of drainage facilities in a UDS is important for reducing flood damage. A preemptive operation is essential for the fast and safe drainage of drainage facilities, such as a CR. The operation of the pump stations at a CR depends on the CR water level, and to achieve a preemptive operation, the inflow directly connected to the water level should be predicted. Additionally, an accurate inflow prediction of a CR is essential for the operation of the pump stations. A multilayer perceptron (MLP) with existing optimizers and combined with metaheuristic optimization algorithms has been suggested to predict the inflow of a CR.
A perceptron, which is a type of artificial neural network (ANN), was developed based on the concept of a hypothetical nervous system and the memory storage of the human brain [1]. The initial perceptron was a single-layer version with the ability to solve only problems that allow linear separations. Hence, an MLP was developed to overcome the abovementioned limitation [2]. Henceforth, various ANN-related approaches, including those pertaining to MLPs, have been proposed. A recurrent neural network (RNN) using error backpropagation was developed for the construction of appropriate internal representations [3]. Additionally, a convolutional neural network (CNN) designed to process two-dimensional shapes was developed for document recognition [4]. Long short-term memory (LSTM), which is a type of RNN, was developed to solve the vanishing gradient problem of a conventional RNN [5]. To solve the shortcomings of LSTM, which requires a significant amount of memory, a gated recurrent unit (GRU) with a fast computational structure simplifying the model structure was proposed [6].
ANNs have been extensively used in investigations pertaining to hydrology and water resources. A multi-output neural network has been recommended for flow duration curve prediction and compared with single-output neural networks [7]. Approaches to the utilization of applicable deep-learning methods for future water resources have also been recommended [8]. Flood susceptibility mapping is a hybrid model that combines a swarm intelligence algorithm with a deep neural network [9]. A standardized streamflow index for hydrological drought using support vector regression, gene expression programming, and M5 model trees has also been suggested [10]. Considering a state-of-the art modeling framework appraisal, artificial intelligence models have been used to predict suspended river sediment transport for future research directions [11]. The spatial pattern of saturated hydraulic conductivity was predicted using a novel genetic algorithm (GA) based hybrid machine learning pedotransfer function [12]. Metaheuristic optimization algorithms, such as the swarm intelligence algorithm, have also been used to improve the performance of an ANN.
Heuristics are simple inference methods that can be promptly used in situations where rational decisions cannot be realized owing to insufficient time or information. It is difficult to develop heuristics when only the characteristics of each problem to be solved are available. Metaheuristics are high-level heuristics that are applicable to various problems without being restricted by the information regarding a specific problem. Although various metaheuristic optimization algorithms exhibit different characteristics, their concept and theory are simple and offer an excellent solution search ability. They can therefore be applied to engineering, sciences, business, and social sciences.
The first metaheuristic optimization algorithm, that is, the genetic algorithm (GA), was proposed [13]. The GA was created by mimicking the crossover and mutation processes in evolutionary theory, after which various notable metaheuristic optimization algorithms were developed, as described in the following. Ant colony optimization is based on the behavior of ants [14]. As an ant identifies a path leading to a food source from its home, it releases pheromones, which rapidly evaporate. However, if the path is optimal, most ants will select it and the pheromone will remain. Particle swarm optimization (PSO) was inspired by the movement of individuals residing in groups, such as ants, fish, and birds. PSO is an algorithm that gradually identifies an optimal solution by comprehensively considering the current velocity of particles and the velocity at which the states of other particles are considered. Harmony search (HS) was developed to encourage improvisation among musicians [15]. The purpose of HS is to identify the best harmony by combining notes produced by each musician. HS has been used extensively owing to its simple structure. Although demonstrating a good performance, the local search for numerical problems is limited when applying such a search. Hence, an improved version called improved harmony search (IHS) has been proposed. IHS, which improves the fine-tuning function based on HS, has demonstrated an excellent performance in various optimization problems [16].
Moreover, hydrological predictions, such as water level and runoff predictions, have been investigated by combining metaheuristic optimization algorithms with MLPs. In fact, researchers have proposed applying the GA as a metaheuristic optimization algorithm in combination with an MLP. Rainfall-runoff modeling was conducted by combining a real-coded genetic algorithm (RCGA) with an MLP [17]. An MLP coupled with an RCGA was suggested for predicting cumulative and discrete rainfall in the Upper Parramatta River basin of Sydney, Australia [18]. An MLP combined with an RCGA was applied to forecast daily rainfall-runoff in the Ourika basin, Morocco [19]. An MLP–RGA hybrid model demonstrated the highest accuracy in water level prediction for the Nakdong River Basin, Korea [20]. In a recent study, HS was applied as a metaheuristic optimization algorithm in combination with an MLP. An MLP combined with HS has also been presented to predict the water level of an urban stream based on pump station discharge [21]. In the abovementioned studies, only the water level or runoff of streams and rivers was predicted. No study has been conducted on the prediction of inflow in urban drainage system. Additionally, minutely prediction for the operation of urban drainage systems were not conducted.
Among non-structural measures, the inflow prediction of a CR for the pump station operation in a UDS is suggested in this study. The inflow prediction of a CR enables the preemptive operation of a pump station. The pump station can be quickly drained based on the inflow prediction of the CR, and it is possible to secure additional storage capacity by maintaining a low CR water level. A new optimizer was proposed by combining IHS, an improved HS that has shown a good performance among various metaheuristic optimization algorithms, with existing optimizers. MLP combined with the new optimizer was applied to the inflow prediction of a CR, the results of which were compared with those of existing optimizers alone and existing optimizers combined with HS.
The characteristics of this study can be divided into four main categories. The first is the selection of monitoring nodes used to consider the state of stormwater conduits. The second is the construction of training data used for an MLP, an MLP combined with HS (MLPHS), and an MLP combined with IHS (MLPIHS). The third is the preprocessing of the training data. The fourth is the inflow prediction of a CR and a comparison of results from each MLP.

2. Methodologies

2.1. Overview

Rainfall, water level data on the first/maximum flooding nodes, and the CR inflow from 2010 to 2020 were acquired. From 2010 to 2019, learning data for rainfall, the water level of the first/maximum flooding nodes, and the CR inflow were collected for the inflow prediction of CR in 2020. After creating the training data, training was conducted by building and applying an MLP, MLPHS, and MLPIHS. The CR inflow was predicted using the rainfall and water level data of the first/maximum flooding nodes as input data. In this study, the applied MLP was constructed using TensorFlow [22]. Figure 1 shows the workflow of this study.

2.2. Preparation of Training Data

The training data must be pre-processed for an accurate prediction of the CR inflow. Although various data preprocessing methods for learning purposes are available, a correlation analysis and normalization were applied in this study.

2.2.1. Correlation Analysis

A correlation analysis was applied to consider the time difference between the input data (rainfall and water level of the first/maximum flooding nodes) and the target data (CR inflow). Equation (1) was used to calculate the correlation coefficient:
r x , y = i = 1 n ( x i x ¯ ) ( y i y ¯ ) i = 1 n ( x i x ¯ ) 2 i = 1 n ( y i y ¯ ) 2  
where rx,y is the correlation coefficient, xi is the input data, yi is the target data, x ¯ is the average of xi, and y ¯ is the average of yi. Figure 2 shows the application of a correlation analysis considering the lag time.

2.2.2. Normalization of Training Data

A study on normalization was conducted when applying ANN to rainfall-runoff models [23]. When the MLP is trained using data with a wide range of values, the prediction performance can degrade owing to the difference between the input and target data. Hence, the data should be converted into values between 0 and 1 through normalization. Normalization was conducted for all data for each input data and target data. The maximum value of each data was converted to 1 and the minimum value of each data was converted to 0. Equation (2) expresses the normalization:
y i = ( x i x a ) ( x b x a )  
where yi is the normalized value, xi is the observed value, xb and xa are the maximum and minimum observed values, respectively.

2.3. MLP Combined with IHS

The MLP is the most basic type of an ANN and comprises one input layer, one or more hidden layers, and one output layer. The weight and bias are set as parameters, and they can be used to express non-linear problems. Figure 3 shows the structure of the MLP including MLPHS and MLPIHS used in this study.
The input and output layers shown in Figure 3, comprise four input data and one output data, respectively. Five hidden layers are used, and each hidden layer contains ten nodes. For an MLP using the existing optimizers, a rectified linear unit (Relu) was applied as the activation function. Including MLPHS and MLPIHS, there were 10,000 epochs for all MLPs.

2.3.1. Existing Optimizers in MLP

The optimizer applied in the MLP was used to compute the weights and biases between nodes. The most basic optimizer is gradient descent (GD), in which a differentiation is conducted to obtain the weights and biases between nodes; however, this entails the use of all available data. Hence, stochastic gradient descent (SGD) was introduced to overcome the shortcomings of GD. Additionally, momentum when considering the inertia was proposed to improve GD, and the Nesterov accelerated gradient (NAG) was suggested to improve the momentum. With SGD, data are randomly selected, whereas with GD, the learning performance differs depending on the learning rate. An adaptive gradient (Adagrad) was proposed to overcome the shortcomings of SGD. With Adagrad, the learning rate can reach zero during the learning process through the application a flexible learning rate based on the learning process, which prevents learning from taking place. To overcome the shortcomings of Adagrad, root mean squared propagation (RMSprop) is an optimizer that improves the learning stoppage using an exponential average, whereas adaptive delta (Adadelta) updates the learning rate using a Hessian matrix and an exponential average. Moreover, the adaptive moment (Adam), a combination of momentum and RMSprop, updates the learning rate using the exponential mean and square of the weights. Adamax, an extension of Adam applying a new infinity norm, has been proposed. Follow the regularized leader (Ftrl), an algorithm that includes normalization of follow the leader by considering the gradient (leader) with the smallest loss, has been proposed. Nadam, which is a combination of NAG and RMSprop, identifies new weights and biases at new locations after propagating in the momentum direction. The existing optimizers applied to the MLP in this study were GD, SGD, Adagrad, RMSprop, Adadelta, Adam, and Nadam. The existing optimizers applied to the MLP in this study were Adadelta, Adagrad, Adam, Adamax, Ftrl, Nadam, RMSprop and SGD. Table 1 shows the existing optimizers in this study.

2.3.2. IHS

HS is a metaheuristic optimization algorithm proposed to encourage improvisation among musicians. The parameters used in HS are the harmony memory size (HMS), harmony memory considering rate (HMCR), pitch adjusting rate (PAR), and bandwidth (BW). HMS refers to the size of the harmony memory (HM) and is the number of candidate solutions that can be stored. HMCR is the probability of randomly selecting decision variables in HM to create a new combination of such variables. PAR refers to the probability of adjusting the decision variable selected through the HMCR when using the BW.
IHS is a metaheuristic optimization algorithm with an improved performance achieved by changing the PAR and BW for a local search in HS based on the number of iterations. In the latter iterations of the metaheuristic optimization algorithm, the effect of a local search, in which a detailed search can be conducted instead of a global search, becomes more important. Equation (3) shows the PAR of IHS:
P A R = P A R m i n + ( P A R m a x P A R m i n ) × ( I c I t )  
where PARmin and PARmax are the lower and upper boundaries of PAR, respectively; Ic is the number of current iterations; and It is the total number of iterations. BW, which is the range of a local search in HS, affects the results. Equation (4) shows the expression of BW in IHS:
  B W = B W m a x + e x p ( ln ( B W m i n B W m a x ) × ( I c I t ) )
where BWmin and BWmax are the lower and upper boundaries of BW, respectively. The IHS procedure is as follows:
  • Step 1. Create initial solutions based on the range of decision variables and generate the HMS.
  • Step 2. Sort the HM in HMS based on the value of the objective function.
  • Step 3(a). Select decision variables in the existing HM when HMCR is applied.
  • Step 3(b). Adjust new decision variables based on BW when PAR is applied.
  • Step 4. Create a new solution using the decision variables created in Steps 3(a) and 3(b).
  • Step 5. Compare the new solution with the worst solution in the existing HM to decide whether it should be replaced.
  • Step 6. Repeat Steps 2 to 5 until the termination criteria is satisfied.
In this study, a sensitivity analysis was conducted to establish the HS parameters. For MLPIHS, a value of 50 was applied for HMS; 0.85, for HMCR; 0.9, for PARmax; 0.1, for PARmin; 0.01, for BWmax; and 0.001, for BWmin.

2.3.3. Combined Optimizer Using Metaheuristic Optimization in MLP

An MLP using HS was recently proposed and applied to predict the runoff in urban streams [21]. Using the MLP proposed herein, a new MLP (MLPIHS) using optimizers combined with IHS has been suggested. Figure 4 shows a calculation flowchart for the MLP using optimizers combined with HS and IHS.

2.4. Selection of Monitoring Nodes

2.4.1. Selection of Maximum Flooding Nodes

The maximum flooding nodes were selected based on the historical rainfall events within the study area. This method is used to find the node with the maximum flooding by applying the rainfall events from 2010 to 2019 as training data. The rainfall-runoff model for each rainfall event was required to obtain the maximum flooding nodes. A rainfall-runoff simulation was conducted using the storm water management model (SWMM) [24]. The selection of the maximum flooding nodes was used in a study on ways to operate drainage facilities for finding the most likely flooding node within the study area [25].

2.4.2. Selection of First Flooding Nodes

The first flooding nodes were selected based on the Huff distribution [26]. The Huff distribution was segmented into four quartiles based on the location of the peak rainfall. The peak value of rainfall in the Huff distribution was located at the 0–25% duration in the first quartile, 25–50% duration in the second quartile, 50–75% duration in the third quartile, and 75–100% duration in the fourth quartile. The third quartile of the Huff distribution, proposed as the appropriate rainfall distribution in Korea, was selected [27]. By applying the third quartile of the Huff distribution, the total amount of rainfall increased by 1 mm, and this process was repeated until the first surcharge/flooding occurred. The simulation was conducted using the storm water management model (SWMM) [24]. Generally, the first flooding occurred between branch conduits; however, because it was difficult to represent the entire drainage system, the first flooding nodes between the main conduits were selected. The branch and main conduits were classified based on the sub-catchment area. If the area of sub-catchment exceeds 0.12 km2, it is classified as a main conduit; otherwise, it is classified as a branch conduit [25]. For the applied duration, the concentration time within the study area was applied up to three times.

3. Application and Results

3.1. Study Area

The Han River passes through the center of Seoul, the capital city of the Republic of Korea, which is used as the study area in this study. The Han River comprises various tributaries, including the Anyang stream. Additionally, the Anyang stream comprises various tributaries, including the Dorim stream. Figure 5 shows the location of the study area and the SWMM network.
Among the 25 districts of Seoul, the Dorim stream flows adjacent to Gwanak-gu, Dongjak-gu, Guro-gu, and Yeongdeungpo-gu. The Dorim stream comprises two tributaries (Daebang and Bongcheon streams) with a length of 11 km and a watershed area of 41.93 km2. Eleven pump stations are installed at the Dorim Stream, and the Daerim3 pump station has the largest drainage area. The Daerim3 pump station comprises 12 pumps (with a drainage capacity of 57.02 m3/s) and has a drainage area of 2.49 km2. The Daerim3 pump station has a CR that can receive inflow from the drainage area, the capacity of which is 36,200 m3. In the network shown on the right side of Figure 5, the CR is located downstream of the drainage area. Additionally, the first and maximum flooding nodes selected in this study are indicated.

3.2. Preparation of Data for Inflow Prediction of CR

Training data must be prepared to predict the inflow of CR. The training data comprise the amount of rainfall, water level of the monitoring nodes, and CR inflow. Data on historical rainfall that occurred within the study area over the past decade were selected. Historical rainfall in this study means that caused flood damage in the study area. Data from 2015 and 2017 were excluded because historical rainfall occurred within the study area. Rainfall records from 2010 to 2020 were obtained from the Korea Meteorological Administration. The CR inflow was obtained based on the water level record of the CR and pump operation records at the Daerim3 pump station. Figure 6 shows the observed rainfall and inflow of the CR within the study area.
Monitoring nodes were selected to verify the status of the urban drainage network. The nodes where the first and maximum flooding occurred were selected as the monitoring nodes. Table 2 shows the results of the maximum flooding nodes when historical rainfall data from 2010 to 2019 were applied.
For all rainfall events, the maximum number of flooding nodes was 550. Node 550 was selected as the monitoring point. For the first flooding nodes, durations of 30, 60, and 90 min were applied, and synthetic rainfall data based on the Huff distribution were used. Table 3 lists the results for the first flooding nodes.
As the results listed in Table 3 indicate, as the first flooding nodes, node 560 was selected at 30 and 60 min, whereas node 575 was selected at 90 min. Nodes 560 and 575 were selected as the monitoring nodes. The water levels of three nodes (550, 560, and 575) were added to the input data. Data on the rainfall, water level of the monitoring nodes, and CR inflow were constructed as the input data.

3.3. Inflow Prediction Using MLPIHS

Before preparing for the training using the input data, a data preprocessing was conducted. A correlation analysis was applied to the input data for the training using MLPs, the results of which are listed in Table 4.
As shown in Table 4, the lag time ranged from 13 to 17 min, which implies that the inflow is predictable after 13 min. Additionally, normalization was conducted to convert all input data for adjusting the data scale. The results of the MLP were obtained by applying Adadelta, Adagrad, Adam, Adamax, Ftrl, Nadam, RMSprop, and SGD provided by TensorFlow as optimizers. The performances of MLPHS and MLPIHS were compared. To calculate the error of each result, the mean square error (MSE) based on the square value was applied along with the mean absolute error (MAE) based on the absolute values. The MSE is expressed as shown in Equation (5):
M S E = i = 1 n ( x o x i ) 2 n
where xo is the observed data, xi is the simulated data, and n is the number of data points. The MAE is expressed as shown in Equation (6):
M A E = i = 1 n | x o x i | n
where xo is the observed data, xi is the simulated data, and n is the number of data points. Table 5 shows the results of the inflow prediction.
Among the MLPs to which the existing optimizers in Table 4 were applied, Adam showed the smallest error in terms of the MSE and MAE. The MSE of all optimizers except for Ftrl, which showed the largest error, did not exceed 10. For MLPHS, Ftrl showed the smallest errors in terms of the MSE and MAE. Ftrl showed a large error for an MLP, whereas the MSE and MAE of Ftrl+HS decreased the most in MLPHS. For MLPIHS, Ftrl+IHS showed the smallest error in terms of the MSE and MAE, and the results were similar to those of Ftrl+HS. Ftrl+IHS showed a slightly improved accuracy compared with Ftrl+HS. Figure 7 shows a comparison of the results of Adam, Ftrl+HS, and Ftrl+IHS.
As shown in Figure 7, the error for Adam was relatively large compared with those of Ftrl+HS and Ftrl+IHS. The difference at the peak value is similar for all three optimizers; however, after the peak value, Ftrl+IHS has a small difference from the observed value.

4. Discussion

The R2 values for the observed and predicted inflows for each MLP were calculated. The value of R2 is expressed as shown in Equation (7):
R 2 = i = 1 n ( x i x ¯ ) 2 i = 1 n ( x o x ¯ ) 2
where xo is the observed data, xi is the simulated data, n is the number of data points, and x ¯ is the average of xo. Additionally, a comparison was conducted based on the flow rates of the observed and predicted inflows. Figure 8 shows a comparison of R2 for the observed and predicted inflows when using Adam.
As shown in Figure 8, the MLP using Adam predicted a slightly higher value at a low inflow and a slightly lower value at a high inflow. Figure 9 shows a comparison of R2 for the observed and predicted inflows when using Ftrl+HS.
As shown in Figure 9, the MLP using Ftrl+HS predicted a slightly higher value at a low inflow and a slightly lower value at a high inflow. With a slight difference, a similar pattern to the results in Figure 8 can be seen. Figure 10 shows a comparison of R2 for the observed and predicted inflows using Ftrl+IHS.
As shown in Figure 10, the MLP using Ftrl+IHS predicted a slightly higher value at a low inflow and a slightly lower value at a high inflow. With a slight difference, a similar pattern to the results shown in Figure 8 and Figure 9 can be seen. Based on the results from Figure 8, Figure 9 and Figure 10, all optimizers tend to predict slightly higher values at a low inflow and lower values at a high inflow. However, the optimizers combining HS and IHS were improved compared with the existing optimizer.
In this study, a structure of MLP that can produce relatively good results was selected through results according to the number of nodes and the number of hidden layers was selected. Figure 11 shows the results according to the number of nodes (Adadelta and Adagrad).
When Adadelta was applied in MLP, the best results were obtained there were 10 nodes. When Adagrad was applied in MLP, the best results were obtained there were 11 nodes. Figure 12 shows the results according to the number of nodes (Adam and Adamax).
When Adam was applied in MLP, the best results were obtained there were 9 nodes. When Adamax was applied in MLP, the best results were obtained there were 12 nodes. Figure 13 shows the results according to the number of nodes (Ftrl and Nadam).
When Ftrl was applied in MLP, the best results were obtained there were 10 nodes. When Nadam was applied in MLP, the best results were obtained there were 8 nodes. Figure 14 shows the results according to the number of nodes (RMSprop and SGD).
When RMSprop was applied in MLP, the best results were obtained there were 10 nodes. When SGD was applied in MLP, the best results were obtained there were 10 nodes. Figure 14 shows the results according to the number of nodes (RMSprop and SGD). Although there were differences for each optimizer, many results showed the smallest MSE when 10 nodes were applied. Figure 15 shows the results according to the number of hidden layers (Adadelta and Adagrad).
When Adadelta was applied in MLP, the best results were obtained there were 7 hidden layers. When Adagrad was applied in MLP, the best results were obtained there were 4 hidden layers. Figure 16 shows the results according to the number of hidden layers (Adam and Adamax).
When Adam was applied in MLP, the best results were obtained there were five hidden layers. When Adamax was applied in MLP, the best results were obtained there were four hidden layers. Figure 17 shows the results according to the number of hidden layers (Ftrl and Nadam).
When Ftrl was applied in MLP, the best results were obtained there were five hidden layers. When Nadam was applied in MLP, the best results were obtained there were seven hidden layers. Figure 18 shows the results according to the number of hidden layers (Ftrl and Nadam).
When RMSprop was applied in MLP, the best results were obtained there were five hidden layers. When SGD was applied in MLP, the best results were obtained there were seven hidden layers. Although there were differences for each optimizer, many results showed the smallest MSE when five hidden layers were applied. Based on the results according to the number of nodes and the number of hidden layers, 10 nodes and five hidden layers were selected.
In the case of a simpler structure of MLP, it could be difficult to show a small MSE value. Figure 19 shows the prediction results according to epochs when Ftrl+IHS is applied.
It is expected that the error could be reduced if training was conducted for more epochs. However, many epochs do not guarantee good results, as overfitting problem could occur.

5. Conclusions

The important aspect of this study is the CR inflow prediction for the preemptive operation of an urban drainage facility such as a pump station. There are two disadvantages to the use of existing MLP optimizers. Existing MLP optimizers can fall into a local optimal solution because they are highly dependent on the initial values (weights and biases). Additionally, because existing optimizers lack storage space, the best result might not be stored during the learning process.
To overcome these two disadvantages, the use of existing optimizers combined with metaheuristic optimization algorithms such as HS and IHS was suggested. Prior to the CR inflow prediction, data preprocessing was conducted by applying a correlation analysis and normalization. To select monitoring nodes as training data, the first flooding nodes and the maximum flooding nodes were obtained. The comparison results showed that the MLP to which the existing optimizers in combination with HS and IHS were applied showed relatively accurate results in terms of both the MSE and MAE. The results revealed that MLPHS and MLPIHS predicted the inflow more accurately than the MLP with existing optimizers.
The model used for inflow prediction in this study runs in a continuous sequence from the beginning. Because prediction was run after training, training and prediction did not proceed simultaneously. However, real-time prediction could be possible if the learning model and the prediction model are executed simultaneously when real-time data is input.
Various follow-up studies can be conducted to overcome this lack of usability. To increase the usability of MLPHS and MLPIHS, an optimized MLP structure can be developed. If metaheuristic optimization algorithms including evolutionary algorithms are applied to optimize the structure (numbers of nodes and hidden layers), it will be possible to improve the usability of an MLP. If the method described in this study is combined with an approach to rainfall prediction, it will be possible to secure additional operational times for urban drainage facilities.

Funding

This study was funded by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (NRF-2019R1I1A3A01059929).

Data Availability Statement

Data available on request due to restrictions (public management policy). The data presented in this study are available on request from the corresponding author with the permission of the public institution. The data are not publicly available due to public management policy.

Acknowledgments

This research was supported by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (NRF-2019R1I1A3A01059929).

Conflicts of Interest

The author declares no conflict of interest.

Abbreviations

CRCentralized reservoir
DRDecentralized reservoir
UDSUrban drainage system
MLPMultilayer perceptron
ANNArtificial neural network
RNNRecurrent neural network
CNNConvolutional neural network
LSTMLong short-term memory
GRUGated recurrent unit
GAGenetic algorithm
PSOParticle swarm optimization
HSHarmony search
IHSImproved harmony search
RCGAReal-coded genetic algorithm
MLPHSMLP using new optimizer combined with HS
MLPIHSMLP using new optimizer combined with IHS
GDGradient descent
SGDStochastic gradient descent
NAGNesterov accelerated gradient
AdagradAdaptive gradient
RMSpropRoot mean squared propagation
AdaDeltaAdaptive delta
AdamAdaptive moment
NadamNesterov accelerated adaptive moment
HMSHarmony memory size
HMCRHarmony memory considering rate
PARPitch adjusting rate
BWBandwidth
HMHarmony memory
SWMMStorm water management model
MSEMean square error
MAEMean absolute error

References

  1. Rosenblatt, F. The perceptron: A probabilistic model for information storage and organization in the brain. Psychol. Rev. 1958, 65, 386. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Minsky, M.; Papert, S. Perceptrons; MIT Press: Cambridge, MA, USA, 1969. [Google Scholar]
  3. Rumelhart, D.E.; Hinton, G.E.; Williams, R.J. Learning representations by back-propagating errors. Nature 1986, 323, 533–536. [Google Scholar] [CrossRef]
  4. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef] [Green Version]
  5. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  6. Chung, J.; Gulcehre, C.; Cho, K.; Bengio, Y. Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling. arXiv 2014, arXiv:1412.3555. [Google Scholar]
  7. Worland, S.C.; Steinschneider, S.; Asquith, W.; Knight, R.; Wieczorek, M. Prediction and inference of flow duration curves using multioutput neural networks. Water Resour. Res. 2019, 55, 6850–6868. [Google Scholar] [CrossRef] [Green Version]
  8. Sit, M.; Demiray, B.Z.; Xiang, Z.; Ewing, G.J.; Sermet, Y.; Demir, I. A comprehensive review of deep learning applications in hydrology and water resources. Water Sci. Technol. 2020, 82, 2635–2670. [Google Scholar] [CrossRef]
  9. Bui, Q.T.; Nguyen, Q.H.; Nguyen, X.L.; Pham, V.D.; Nguyen, H.D.; Pham, V.M. Verification of novel integrations of swarm intelligence algorithms into deep learning neural network for flood susceptibility mapping. J. Hydrol. 2020, 581, 124379. [Google Scholar] [CrossRef]
  10. Shamshirband, S.; Hashemi, S.; Salimi, H.; Samadianfard, S.; Asadi, E.; Shadkani, S.; Kargar, K.; Mosavi, A.; Nabipour, N.; Chau, K.W. Predicting standardized streamflow index for hydrological drought using machine learning models. Eng. Appl. Comput. Fluid Mech. 2020, 14, 339–350. [Google Scholar] [CrossRef]
  11. Tao, H.; Al-Khafaji, Z.S.; Qi, C.; Zounemat-Kermani, M.; Kisi, O.; Tiyasha, T.; Chau, K.; Nourani, V.; Melesse, A.M.; Elhakeem, M.; et al. Artificial intelligence models for suspended river sediment prediction: State-of-the art, modeling framework appraisal, and proposed future research directions. Eng. Appl. Comput. Fluid Mech. 2021, 15, 1585–1612. [Google Scholar] [CrossRef]
  12. Singh, V.K.; Panda, K.C.; Sagar, A.; Al-Ansari, N.; Duan, H.F.; Paramaguru, P.K.; Vishwakarma, D.K.; Kumar, A.; Kumar, D.; Kashyap, P.S.; et al. Novel Genetic Algorithm (GA) based hybrid machine learning-pedotransfer Function (ML-PTF) for prediction of spatial pattern of saturated hydraulic conductivity. Eng. Appl. Comput. Fluid Mech. 2022, 16, 1082–1099. [Google Scholar] [CrossRef]
  13. Goldberg, D.E.; Holland, J.H. Genetic algorithms and machine learning. Mach. Learn. 1988, 3, 95–99. [Google Scholar] [CrossRef]
  14. Dorigo, M. Optimization, Learning and Natural Algorithms. Ph.D. Thesis, Politecnico di Milano, Milan, Italy, 1992. [Google Scholar]
  15. Geem, Z.W.; Kim, J.H.; Loganathan, G.V. A new heuristic optimization algorithm: Harmony search. Simulation 2001, 76, 60–68. [Google Scholar] [CrossRef]
  16. Mahdavi, M.; Fesanghary, M.; Damangir, E. An improved harmony search algorithm for solving optimization problems. Appl. Math. Comput. 2007, 188, 1567–1579. [Google Scholar] [CrossRef]
  17. Srinivasulu, S.; Jain, A. A comparative analysis of training methods for artificial neural network rainfall–runoff models. Appl. Soft Comput. 2006, 6, 295–306. [Google Scholar] [CrossRef]
  18. Nasseri, M.; Asghari, K.; Abedini, M.J. Optimized scenario for rainfall forecasting using genetic algorithm coupled with artificial neural network. Expert Syst. Appl. 2008, 35, 1415–1421. [Google Scholar] [CrossRef]
  19. Sedki, A.; Ouazar, D.; El Mazoudi, E. Evolving neural network using real coded genetic algorithm for daily rainfall–runoff forecasting. Expert Syst. Appl. 2009, 36, 4523–4527. [Google Scholar] [CrossRef]
  20. Yeo, W.K.; Seo, Y.M.; Lee, S.Y.; Jee, H.K. Study on water stage prediction using hybrid model of artificial neural network and genetic algorithm. J. Korea Water Resour. Assoc. 2010, 43, 721–731. [Google Scholar]
  21. Lee, W.J.; Lee, E.H. Runoff prediction based on the discharge of pump stations in an urban stream using a modified multi-layer perceptron combined with meta-heuristic optimization. Water 2022, 14, 99. [Google Scholar] [CrossRef]
  22. Google Research. Tensorflow Serving. 2016. Available online: www.tensorflow.org/tfx/guide/serving?hl=ko (accessed on 10 January 2023).
  23. Minns, A.W.; Hall, M.J. Artificial neural networks as rainfall-runoff models. Hydrol. Sci. J. 1996, 41, 399–417. [Google Scholar] [CrossRef]
  24. United States Environmental Protection Agency. Storm Water Management Model User’s Manual Version 5.0; EPA: Washington DC, USA, 2010. [Google Scholar]
  25. Lee, E.H.; Lee, Y.S.; Joo, J.G.; Jung, D.; Kim, J.H. Investigating the impact of proactive pump operation and capacity expansion on urban drainage system resilience. J. Water Resour. Plan. Manag. 2017, 143, 04017024. [Google Scholar] [CrossRef]
  26. Huff, F.A. Time distribution of rainfall in heavy storms. Water Resour. Res. 1967, 3, 1007–1019. [Google Scholar] [CrossRef]
  27. Yoon, Y.N.; Jung, J.H.; Ryu, J.H. Introduction of design flood estimation. J. Korea Water Resour. Assoc. 2013, 46, 55–68. [Google Scholar]
Figure 1. Workflow of this study.
Figure 1. Workflow of this study.
Water 15 01543 g001
Figure 2. Application of a correlation analysis considering the lag time.
Figure 2. Application of a correlation analysis considering the lag time.
Water 15 01543 g002
Figure 3. Structure of the MLP including MLPHS and MLPIHS used in this study.
Figure 3. Structure of the MLP including MLPHS and MLPIHS used in this study.
Water 15 01543 g003
Figure 4. Calculation flowchart for MLP using optimizers combined with HS and IHS.
Figure 4. Calculation flowchart for MLP using optimizers combined with HS and IHS.
Water 15 01543 g004
Figure 5. Location of study area and network of SWMM.
Figure 5. Location of study area and network of SWMM.
Water 15 01543 g005
Figure 6. Observed rainfall and inflow of CR within the study area.
Figure 6. Observed rainfall and inflow of CR within the study area.
Water 15 01543 g006
Figure 7. Comparison of prediction for each method.
Figure 7. Comparison of prediction for each method.
Water 15 01543 g007
Figure 8. Comparison of R2 for observed and predicted inflows when using Adam.
Figure 8. Comparison of R2 for observed and predicted inflows when using Adam.
Water 15 01543 g008
Figure 9. Comparison of R2 for observed and predicted inflows when using Ftrl+HS.
Figure 9. Comparison of R2 for observed and predicted inflows when using Ftrl+HS.
Water 15 01543 g009
Figure 10. Comparison of R2 for observed and predicted inflows when using Ftrl+IHS.
Figure 10. Comparison of R2 for observed and predicted inflows when using Ftrl+IHS.
Water 15 01543 g010
Figure 11. Results according to the number of nodes (Adadelta and Adagrad).
Figure 11. Results according to the number of nodes (Adadelta and Adagrad).
Water 15 01543 g011
Figure 12. Results according to the number of nodes (Adam and Adamax).
Figure 12. Results according to the number of nodes (Adam and Adamax).
Water 15 01543 g012
Figure 13. Results according to the number of nodes (Ftrl and Nadam).
Figure 13. Results according to the number of nodes (Ftrl and Nadam).
Water 15 01543 g013
Figure 14. Results according to the number of nodes (RMSprop and SGD).
Figure 14. Results according to the number of nodes (RMSprop and SGD).
Water 15 01543 g014
Figure 15. Results according to the number of hidden layers (Adadelta and Adagrad).
Figure 15. Results according to the number of hidden layers (Adadelta and Adagrad).
Water 15 01543 g015
Figure 16. Results according to the number of hidden layers (Adam and Adamax).
Figure 16. Results according to the number of hidden layers (Adam and Adamax).
Water 15 01543 g016
Figure 17. Results according to the number of hidden layers (Ftrl and Nadam).
Figure 17. Results according to the number of hidden layers (Ftrl and Nadam).
Water 15 01543 g017
Figure 18. Results according to the number of hidden layers (RMSprop and SGD).
Figure 18. Results according to the number of hidden layers (RMSprop and SGD).
Water 15 01543 g018
Figure 19. Prediction results according to the epochs when Ftrl+IHS is applied.
Figure 19. Prediction results according to the epochs when Ftrl+IHS is applied.
Water 15 01543 g019
Table 1. Existing optimizers in this study.
Table 1. Existing optimizers in this study.
OptimizersDescription
AdadeltaUpdate the learning rate using a Hessian matrix and exponential average
AdagradUse a flexible learning rate based on the learning process
AdamCombine momentum and RMSporp
AdamaxApply a new infinity norm
FtrlNormalize the follow the leader by considering gradient (leader) with the smallest loss
NadamCombine Nesterov accelerated gradient and RMSprop
RMSpropImprove learning stopping using an exponential average
SGDSelect randomly from the entire data set
Table 2. Results of maximum flooding nodes.
Table 2. Results of maximum flooding nodes.
Rainfall Events20102011201220132014201620182019
Maximum flooding nodes550550550550550550550550
Table 3. Results of first flooding nodes.
Table 3. Results of first flooding nodes.
Duration (m)306090
First flooding nodes560560575
Table 4. Results of correlation analysis.
Table 4. Results of correlation analysis.
Data TypeMonitoring Node (550)Monitoring Node (560)Monitoring Node (575)Rainfall Data
Lag time (min)15141317
Correlation coefficient0.8130.9490.9520.747
Table 5. Results of inflow prediction.
Table 5. Results of inflow prediction.
MethodAdadeltaAdagradAdamAdamaxFtrlNadamRMSpropSGD
MSE4.3120005.9700603.0829333.10690111.9787073.1992553.2617816.689764
MAE1.2857711.2611581.0213781.0244761.6351111.0366191.0856431.443204
MethodAdadelta

+HS
Adagrad
+HS
Adam
+HS
Adamax
+HS
Ftrl
+HS
Nadam
+HS
RMSprop
+HS
SGD
+HS
MSE3.1992553.0885223.0715383.0784172.9369293.1370133.0343064.901860
MAE1.0366191.0288091.0189221.0251791.0017931.0449881.0153441.304466
MethodAdadelta
+IHS
Adagrad
+IHS
Adam
+IHS
Adamax
+IHS
Ftrl
+IHS
Nadam
+IHS
RMSprop
+IHS
SGD
+IHS
MSE3.0293702.9842123.0470003.0466822.9301923.1198313.0203943.247332
MAE1.0179071.0204831.0308211.0247050.9895321.0375071.0001691.177632
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lee, E.H. Inflow Prediction of Centralized Reservoir for the Operation of Pump Station in Urban Drainage Systems Using Improved Multilayer Perceptron Using Existing Optimizers Combined with Metaheuristic Optimization Algorithms. Water 2023, 15, 1543. https://doi.org/10.3390/w15081543

AMA Style

Lee EH. Inflow Prediction of Centralized Reservoir for the Operation of Pump Station in Urban Drainage Systems Using Improved Multilayer Perceptron Using Existing Optimizers Combined with Metaheuristic Optimization Algorithms. Water. 2023; 15(8):1543. https://doi.org/10.3390/w15081543

Chicago/Turabian Style

Lee, Eui Hoon. 2023. "Inflow Prediction of Centralized Reservoir for the Operation of Pump Station in Urban Drainage Systems Using Improved Multilayer Perceptron Using Existing Optimizers Combined with Metaheuristic Optimization Algorithms" Water 15, no. 8: 1543. https://doi.org/10.3390/w15081543

APA Style

Lee, E. H. (2023). Inflow Prediction of Centralized Reservoir for the Operation of Pump Station in Urban Drainage Systems Using Improved Multilayer Perceptron Using Existing Optimizers Combined with Metaheuristic Optimization Algorithms. Water, 15(8), 1543. https://doi.org/10.3390/w15081543

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop