Next Article in Journal
Chironomid Pupal Exuviae Technique in Ecological Research of Man-Made Water Bodies
Previous Article in Journal
Exploring the Drivers Influencing Multidimensional Alpha and Beta Diversity of Macroinvertebrates in Mountain Streams
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Echo State Network and Sparrow Search: Echo State Network for Modeling the Monthly River Discharge of the Biggest River in Buzău County, Romania

1
National Key Laboratory of Deep Oil and Gas and School of Geosciences, China University of Petroleum (East China), Qingdao 266580, China
2
Department of Civil Engineering, Transilvania University of Brasov, 5, Turnului Street, 500152 Brasov, Romania
*
Author to whom correspondence should be addressed.
Water 2024, 16(20), 2916; https://doi.org/10.3390/w16202916
Submission received: 10 August 2024 / Revised: 26 September 2024 / Accepted: 12 October 2024 / Published: 14 October 2024
(This article belongs to the Section Hydrology)

Abstract

:
Artificial intelligence (AI) has become an instrument used in all domains with good results. The water resources management field is not an exception. Therefore, in this article, we propose two machine learning (ML) techniques—an echo state network (ESN) and sparrow search algorithm–echo state network (SSA-ESN)—for monthly modeling of the water discharge of one of the biggest rivers in Romania for three periods (S, S1, and S2). In both models, R2 was over 0.989 on the test and training sets and the mean absolute error (MAE) varied between 4.4826 and 7.6038. The performance of the SSA-ESN was similar, but the ESN had the shortest run time. The influence of anomalies on the models’ quality was assessed by running the algorithms on a series without the aberrant values, which were detected by the seasonal hybrid extreme studentized deviate (S-H-ESD) test. The results indicate that removing the anomalies significantly improved both models’ performance, but the run time was increased.

1. Introduction

In recent decades, the demand for metaheuristic algorithms has surged alongside the growing complexity of various real-life problems. While traditional optimization methods struggle with the exponential growth in search space, metaheuristics offer efficient and effective solutions. Given that the constraints (such as convexity and nonlinearity, autocorrelation, normality, etc.) that should be fulfilled by variables when using classical optimization approaches are avoided when employing machine learning (ML) algorithms, these algorithms represent a change in perspective and a significant advancement in addressing optimization challenges [1,2].
In the case of water resources management and hydraulics, the advantage of sophisticated computational models aimed at enhancing streamflow forecast precision has the potential to yield substantial cost savings. Various methods have been used for this purpose. For example, Tanty and Desmukh [3] presented some applications of artificial neural networks in hydrology, and Li and Young [4] modeled a suspended sediment load by machine learning. Hayder et al. [5] used PSO–FNN in a study of river discharge in Malaysia, whereas Khan et al. [6] predicted the water level of a river in India. To predict the dynamic pressure distribution in hydraulic structures, Samadi et al. [7] investigated the performances of gene expression programming, artificial neural networks (ANNs), and classification and regression trees. Haghiabi et al. [8] compared the efficiency of a multilayer perceptron and adaptive neuro-fuzzy inference system in predicting triangular labyrinth weirs’ discharge coefficient and found that the first technique gave the best results. Other authors [9,10,11,12] studied ultrasound cavitation’s effect in water by AI techniques.
In the context of forecasting the water flow of a river in Pakistan, Adnan et al. [13] demonstrated the efficiency and reliability of ML techniques. They combined locally weighted learning (LWL) with additive regression (AR), dagging (DG), bagging (BG), random subspace (RS), and rotation forest (RF) to achieve accurate predictions. Van Thieu et al. [14] integrated extreme learning machines with three AI techniques to improve the streamflow forecast of the Nile River (at Aswan). Other studies have also confirmed the effectiveness of ML techniques in predicting rivers’ water discharge based solely on historical records and assessing flood susceptibility in different catchments [15,16,17,18,19,20,21,22,23].
ESN falls in to the recurrent neural network (RNN) category. It functions by inputting data, followed by triggering a nonlinear response in the neurons from the “reservoir” (hidden layer). ESN output consists of a linear combination (trained for optimizing the weights) of all neurons’ responses in the hidden layer. This network mitigates gradient explosion/vanishing by addressing RNN weight selection [24]. Given its ability to handle complex data, with high variability, robust generalization capabilities, and the use of inexpensive training procedure, ESN has been used in various domains [25], such as time-series analysis, forecast, and classification [26,27], signal processing [28], speech recognition [29,30], pattern classification [31,32], finance, bioinformatics, and engineering [33,34,35,36,37], etc. It was also proved that ESN has a low sensitivity to parameter variation [33]. Compared to other algorithms, such as BP and RNN, ESN offers superior prediction accuracy, faster computation, good generalization capabilities, and robust nonlinear mapping capabilities [36,37].
The advantages of ESN for time-series forecasting are: (1) the reservoirs’ weights are fixed during training after being randomly generated [27], (2) the memoryless readout network that allows linear regression for training the weights [28], (3) the short-term memory [36], and (4) the fact that computation of the derivatives is not necessary for training [38].
Introduced in 2020, SSA, based on sparrows’ foraging and anti-predation behavior, has quickly gained popularity due to its simplicity and efficiency in global optimum detection. Since its inception, SSA has been used in various research fields, such as civil [39], electrical [40,41], industrial [42], and mechanical engineering [43], image processing [44], networking [45], communication [46], environmental modeling [47,48], with remarkable results. Several extended versions of SSA, including hybridization, have also been proposed to prevent premature convergence and enhance diversity [49].
The Buzău River, studied in this article, is one of the main rivers in Romania and on which one of the biggest dams was built. Only a few studies on its water discharge have been performed. Modeling this data series is challenging due its statistical features [50,51,52]. In [50], the authors analyzed the minima, maxima, and standard deviation of the Buzău monthly river flow to assess the data homogeneity of the series. The output indicated the necessity of a deeper investigation in this direction. The Mann–Kendall test and the non-parametric Sen slope [51,52] at a significance level of 5% indicated an increasing trend in annual, monthly, fall, and winter series after the dam had been put into operation. However, the fit trend was linear and could not capture the series oscillations.
Another attempt to evaluate the dam’s impact on the water discharge was made in [53] using IHA indicators and software. The output showed the extent to which the indicators changed after 1984 with respect to the previous period. Two parametric regression models of the recorded flow at Nehoiu vs. Bâsca (the main affluent of Buzău upstream Nehoiu) were also proposed. Generalized regression neural network models trained on 1955–1983 and tested on 1984–2010 series were not satisfactory [54]. Three articles [55,56,57] provided promising results in modeling the studied series using AI methods. Within this framework, the primary purpose of this article is to build forecasting models for the monthly river discharge of the Buzău River for the periods 1955–2010, 1955–1983, and 1984–2010 using ESN and SSA-ESN. These algorithms were chosen because despite SSA and ESN’s confirmed performances in various applications, ESN was rarely employed for hydrological predictions. Since our scientific literature search did not find any applications of the hybrid SSA-ESN in this field, this study investigates the suitability of ESN and SSA-ESN for forecasting long series of hydrological records. It also proves the superiority of this approach against others, such as ARIMA, MLP, BPNN, ELM, ESN, LSTM, CNN-LSTM, CNN-LSTM, and SSA-BP on three series.
Another aspect rarely discussed in hydro-meteorological data modeling is the influence of aberrant values on the model’s quality. To our knowledge, such an analysis has not been performed for ESN or SSA-ESN models built on any kind of series. This paper provides a comparison between the output obtained after running these algorithms on raw series and datasets without outliers using three goodness-of-fit indicators. Moreover, comparisons of run time in both scenarios indicate an increase in run time when aberrant values are removed from the series and open the direction for a future investigation of this behavior of the algorithms.
Given its specific design, this study emphasizes that the river flow pattern changed after 1984, as reflected by the performances of the models proposed for the entire series and its subseries.

2. Data Series and Methodology

2.1. Study Area and Data Series

With a catchment of 5264 km2 situated in the Curvature Carpathians, a 1043 m mean elevation, and a length of 302 km, the Buzău River is a tributary of the Siret, located in southeastern Romania (Figure 1).
In the mountainous sector, the river’s dynamic is significant, mainly when increased precipitations occur. The climate in Buzău County is temperate continental, varying with the altitude, relief orientation, and configuration from north to south. The most important tributaries of the river in the mountainous zone and sub-Carpathians, such as Bâsca Roziliei, Bâsca Chiojdului, Pănătău, Slănic, and Câlnău, play a crucial role in the region’s water system. Upstream of Nehoiu, the river collects over 80% of the annual water volume. Due to the building of the Siriu Dam, the Buzău River discharge was altered with respect to the period before January 1984 [58,59].
The analyzed series, denoted in the following by S, consists of the monthly water discharge recorded from the beginning of 1955 to the end of 2010. The series was provided by the National Institute of Hydrology and Water Administration, where the raw datasets were meticulously preprocessed to remove any measurement errors, ensuring a level of data accuracy.
Generally, various techniques can be used to remove inconsistencies and/or incoherencies in datasets. Homogenization is such a solution when a significant number of series recorded in the same region on various catchments are available [60,61,62,63]. When a singular data series is available, removing inconsistencies from the data series can be achieved by singular spectrum analysis [64,65,66], which is based on the reconstruction of a data series after a specific type of decomposition. Another possible approach involves other kinds of decompositions, such as empirical mode decomposition (EMD) and ensemble empirical mode decomposition (EEMD) [67,68,69,70]. A deeper investigation should be conducted in a future study.
The data series was complete. In an opposite situation, different approaches to missing data are possible: (1) to let the algorithm run (likely because it works with labeled values, in contrast with the classical methods for time-series modeling) and (2) yo replace the absent values with the series mean or median, values resulting from linear interpolation, or generated by AI algorithms called autoencoders.
Two subseries of S were also studied: S1—before January 1984 and S2—after December 1983. Figure 2 depicts the series and the values of some statistics. Min (m3/s) is the minimum, Max (m3/s) is the maximum, and CV (%) is the coefficient of variance. Skewness and Kurtosis are dimensionless. Excepting Min, all the basic statistics corresponding to S2 are the lowest, indicating an attenuation of the river flow after 1984.

2.2. Methodology

To better understand the characteristics of the series S, apart from the basic statistics (displayed in Figure 2), the series was tested for stationarity against nonstationarity with the KPSS test [71], homoscedasticity vs. heteroskedasticity by the Fligner–Killeen test [72], and change point existence against nonexistence by the Pettitt test [73].
Anomalies (outliers) are records that significantly differ from the general behavior of a time series and cannot be considered measurement errors or noise [74]. Identifying anomalies in a time-series dataset is crucial because they may impact forecast model accuracy. Even a small number of outliers can considerably diminish the reliability and precision of the forecasts.
The procedures developed for detecting anomalies in the temporal domain take into account the local or global aspect. In the first case, the vicinities of each record are considered, whereas in the second one, the entire series is analyzed [75]. In this research, we utilized a G-H-ESD procedure for detecting multiple outliers, implemented in the R package AnomalyDetection [76], based on the generalized extreme studentized deviate proposed by Roesner [77]. The algorithm’s input is formed by the time series, its length (n), and the maximum number of anomalies, k ≤ 0.49 × n.
The stages of the algorithm are as follows [78].
(a)
Determine the seasonal component (St) using the seasonal and trend decomposition using LOESS (STL) [79,80]. STL consists of two loops: the inner one deriving the trend (Tt), St, and residual (Rt), and the outer one responsible for the algorithm’s robustness with respect to anomalies.
(b)
Compute the median of the given series.
(c)
Compute the residual by subtracting St and the median from the data series.
(d)
Detect the anomalies using ESD as follows.
  • Compute
    C j = max 1 j n x j x ¯ / s ,
    for the extreme value x j detected, where x ¯ = average and s = standard deviation.
  • Compare Cj with the critical value:
    λ j = ( n j ) t p , n j 1 ( n j 1 + t p , n j 1 2 ) ( n j + 1 ) ,
    where t p , n j 1 is the value of the Student statistics at the significance level p and n j 1 degrees of freedom.
  • If xj is an anomaly, discard it and compute the critical values using the new data series.
  • Repeat the previous steps j times, considering that the number of anomalies is equal to the highest j for which Cj > λj.
(e)
List the anomalies and the corresponding timestamp.
The series S, S1, S2, and those obtained after the anomalies removal, denoted by S_a, S1_a, S2_a, are modeled using ESN and SSA-ESN.
When implementing AI methods, it is customary to partition the analyzed series into two disjoint parts, the first to train the model and the other to validate it. The series partition can significantly influence the prediction accuracy. Hence, carefully considering the proportion between these two sets is necessary. A higher proportion of data in the training set can help the model recognize patterns within it, while a lower proportion might lead to an incomplete evaluation of the model’s effectiveness.
The series were standardized before modeling. The training sets were before January 2006 for S, before January 1984 for S1, and between January 1984 and December 2005 for S2. For comparison reasons, the test set was kept unchanged, i.e., January 2006–December 2010. The same sets were kept for modeling S_a, S1_a, S2_a.
Remember that the choice of the random seed plays a significant role in ML algorithms, influencing the weight initialization and the selection of data involved in the algorithm’s deployment. To ensure that results can be reproduced, it is necessary to fix the seed. To evaluate the models’ performance, the algorithms were executed varying the seeds and then choosing the best results.
The models’ quality was assessed employing MAE, MSE, and R2 for the training and test sets. Smaller MAE and MSE or bigger R2 values indicate better models.
The technical specifications of the hardware and software used in the evaluation process are of utmost importance. The hardware comprised an AMD Ryzen 9 5900X 12-core processor CPU (3.70 GHz, 12 cores, 24 threads) with 64 GB of RAM, and an NVIDIA GeForce RTX 3090 GPU. Matlab R2023a was used to implement the models.

2.2.1. ESN

ESN [38] represents a version of a recurrent neural network [81], designed for tasks that require sequential data processing. Its structure, displayed in Figure 3, is optimized for this purpose.
ESN is formed by the input, hidden (reservoir), and output layers. Consider an ESN with k, n, and m input, reserve, and output units. The input, reservoir, and output matrices are U, X, and Y, respectively. The matrix of the connection weights between the first and second layers ( W i n R n × k ) and that of the connections inside the second layer ( W R n × k ) are randomly generated and remain unchanged. To achieve predictive output is sufficient to train the connection weights between the last two layers included in the matrix ( W o u t R m × ( k + n ) ) [82,83].
At stage t, the reservoir layer’s state requires an update, which is described by Equation (3):
X t + 1 = f ( W i n · U t + 1 + W · X t + W b a c k · Y ( t ) )
where f is the reservoir layer neurons’ activation function (usually sigmoid or tanh). W b a c k is the feedback matrix, which is null when the user does not require the output feedback.
The output state equation of the ESN model is:
Y t + 1 = W o u t z t + 1 ,
where z ( t + 1 ) is the concatenation of X t + 1 and U t + 1 [24].
Due to the RNN’s auto-feedback property, X t reflects the previous input traces, so ESN has a short-term memory that dynamically changes.
The parameters utilized for running the following are as follows:
  • Number of samples 30;
  • Number of neurons in the reservoir 1000;
  • Learning rate 0.1;
  • Regularization parameter 0.1 [21].

2.2.2. SSA-ESN

SSA [84] is a novel metaheuristic algorithm based on the simulation of the sparrow flock foraging: sparrows that find better food act as “discoverers,” while others become “followers.” Additionally, the algorithm selects a certain proportion of sparrows for scouting and early warning. Upon detecting danger, they immediately abandon food to prioritize safety. The following matrix presents sparrow positions:
Y = y 1 , 1 y 1 , 2 y 1 , d y 2 , 1 x 2 , 2 y 2 , d y n , 1 y n , 2 y n , d ,
n being the volume of the population and d the dimension of the variable. The fitness function of the sparrow flock is given by the following:
F Y = f ( [ y 1 , 1 y 1 , 2 y 1 , d ) ] f ( [ y 2 , 1 y 2 , 2 y 2 , d ) ] f ( [ y n , 1 y n , 2 y n , d ) ] ,
where the fitness score of each sparrow is found in a row of the matrix F and represents the energy reserve.
Discoverers with higher energy reserves are prioritized for food acquisition during the search process. Discoverers typically have a broader search range than followers, comprising 10% to 20% of the population. The update rule (at the moment t + 1) for the positions of discoverers is as follows:
Y i , k t + 1 = Y i , k t · e x p i α · i t e r m a x , i f   R 2 < S T Y i , k t + Q · L , i f   R 2 S T ,
α 0,1 ,   R 2 0,1 , S T 0.5,1 .
In (7), i t e r m a x = the highest number of iterations, Y i , k = information on the position of the i-th individual in the flock in the dimension k, and t = the iteration moment. The numbers α and Q are random, the latter being generated from a Gaussian distribution. R 2 and S T are the warning and safety values, respectively, and L is a row vector with d elements equal to 1.
Equation (8) describes the process of updating the followers’ (joiners’) positions:
Y i , k t + 1 =   Q · e x p Y w o r s t t Y i , k t k 2 ,   i f   k > n / 2 Y P t + Y i , k t Y P t + 1 · A + · L ,   o t h e r w i s e
where Y p is the actual best position of a discoverer, Y w o r s t is the actual global worst location, “A” is a row matrix where each element is either 1 or −1, “⋅” indicates element-wise multiplication, and:
A + = A T A A T 1 .
When danger is detected, the sparrow flock acts according to the following rules in the case of danger:
Y i , k t + 1 = Y b e s t t · β · Y i , k t Y b e s t t + 1 ,   i f   f i > f g Y i , k t + M · Y i , k t Y w o r s t t f i f g + β ,   i f   f i = f g
where Y b e s t is the actual global best location, β controls the step size, and M ∈ [−1,1]. f i denotes the i-th individual current fitness value, f g and f w are the current global and worst best, and β is a very small number added to prevent division by zero.
The SSA steps for optimizing ESN reservoir parameters are as follows.
(1)
Data preprocessing: Normalize the input time-series data to eliminate scale differences, enhance model convergence speed, and improve prediction accuracy.
(2)
Parameter initialization: Set the SSA’s key parameters—the size of the sparrow population, scouting and warning rate, flight distance (R2)—and the ESN’s basic parameter ranges (such as reservoir size, initial state, spectral radius, input weight).
(3)
ESN parameter optimization: Based on SSA’s randomly generated positions, calculate the population fitness according to the update formula and obtain the current global optimum and individual best values.
(4)
Iteration termination: If the criteria for stopping the iterations are satisfied, the iteration is stopped and the optimal result is listed. Otherwise, the algorithm is performed again from the third step for further iteration.
(5)
ESN network prediction: Select the best individual from SSA as the optimization solution for reservoir parameters. Utilize these optimal parameters for ESN model prediction.
The parameters used to run the hybrid ESN-SSA are the following:
  • Number of parameters to be optimized = 3—learning rate, reservoir size, regularization coefficient.
  • Lower bounds for the parameters—0.1, 100, and 0.1, respectively.
  • Lower bounds for the parameters—2000, 1500, and 0.2, respectively.
  • Sparrow population—10.
  • Maximum number of iterations—50.
  • Initial size of the reservoir—30.

3. Results and Discussion

The KPSS test did not reject the hypothesis of level stationarity, but did reject that of trend stationarity. The Fligner–Killeen test for homogeneity of variances rejected the series’ homoscedasticity (p-value = 0.001385). The Pettitt breakpoint test indicates the existence of a breakpoint in the time series. Moreover, 29 aberrant values were found by the statistical tests.
After running S-H-ESD, 29 anomalies were determined. Figure 4a presents the series and the anomalies (blue dots) and lists the moment of their apparition. Eighteen of them were recorded before January 1984, and most were in March–May. The series without anomalies is presented in Figure 4b.
Figure 5 contains the charts of the S, S1, and S2 (in blue) and the predicted series obtained by ESN (in red) for the test period January 2006–December 2010. It shows that the most significant differences between the raw series and the forecast are registered for the maxima and minima of the test set, meaning that the extreme values in the test set are not very well estimated.
Comparisons of Figure 5a–c indicate that the best evaluation of the minima is provided for S2, especially for the periods May 2007–June 2007, November 2007–December 2007, and February 2010–September 2010. No significant difference is noticed for the maxima. All models follow exactly the data series shape.
The quality of the ESN output was assessed by computing the mean absolute error (MAE), mean standard error (MSE), and R2. Their values are contained in Table 1. The run time was under 1 s, proportional to the series length: the shortest for S2 and the longest for S.
In terms of MAE and MSE, the best model was that for S2, followed by that for S, whereas with respect to R2, the ESN model for S is followed by that for S2. The ESN model for S1 was the worst for both training and test sets. In all cases, the MSE (MAE) for the test set was at least 1.56 (1.26) times lower than for the training test, indicating that the algorithm applies well to the test set what it learned from the training set. The values of R2 were comparable on both sets (test and training).
The goodness-of-fit indicators for the hybrid SSA-ESN models are presented in Table 2. The run time is lower in all situations compared to the ESN. The other indicators are almost the same as in Table 1, suggesting that the initial algorithm (ESN) returned optimized results.
The explanation for the S2 models being the best is that the training and test set belong to the same period (after January 1984). Model S is built on a training set that contains elements before and after building the dam, while the test set belongs to the last period. Hence, its accuracy is lower than that of the model for S. The worst model is that for S1, whose training set is before building the dam, and its test set is after that date. This result indicates a modification in the water flow regime after January 1984.
Table 3 and Table 4 present the models’ goodness-of-fit indicators, providing a clear comparison of their performance.
Figure 6, on the other hand, offers a visual representation of the test set and the corresponding forecasts. Comparison of the ESN and SSA_ESN on the series S_a, S1_a, S2_a leads to the following conclusions.
  • The run time of SSA_ESN is much higher than that of ESN. Indeed, for the hybrid algorithm (Table 4), it is 155.59 s (62.99 s and 110.31 s) for S_a, (S1_a and S2_a), whereas for the single one it is 2.27 s (1.06 s, and 0.94 s) for the same sets (Table 3).
  • Since MAE from the hybrid algorithm belongs to the interval [4.47, 4.86] for the test and [4.16, 4.42] for the training set, compared to the intervals [4.98, 5.1] and [4.40, 4.87], respectively, it results that SSA-ESN performs better in terms of MAE.
  • With respect to MSE, ESN is the best on the training set for S_a and S2_a (40.33 and 39.33—Table 3, compared to 44.49 and 39.73—Table 4), and on the test set for S_a, S1_a, and S2_a (33.28, 37.95, and 32.21, respectively—Table 3, compared to 37.42, 37.97, and 35.46, respectively—Table 4).
  • Considering R2, ESN is the best on all series after discarding the aberrant values compared to SSA-ESN.
Overall, the following conclusions can be drawn.
  • The lowest run time was that of ESN on S2.
  • The lowest MAEs were recorded for SSA-ESN on S2, with values of 4.16 on the training set and 4.47 on the test set.
  • The lowest MSEs were obtained by running ESN on S2_a: 39.33 on the training set and 32.21 on the test set.
  • The highest R2 (over 99.99%) corresponds to ESN on S2_a.
  • The removal of aberrant values significantly enhanced the performance of the ESN algorithm, demonstrating its adaptability.
We have to point out that in this article, the test set was 60 months in all cases, so the forecast period was kept at 60 months for comparison reasons. The analysis can be conducted for various training:test ratio, so for different forecast periods (longer or shorter than 60 months). Obviously, the algorithms’ performance will be different. Such a study will be conducted in the future.
Modeling hydrological data is challenging due to the high variability of river flow, reflected in the values of the basic statistics, stationarity, change point existence, and heteroskedasticity, as in the present case study. Under these conditions, one cannot suppose that elementary methods will yield good results in capturing all these features. Still, we fitted ARIMA models [56] to compare their performances with those of AI models. The results of the best-fit ARIMA model will be presented in the following.
Another issue is that a wide range of articles should report the use of ESNs in modeling hydrological data to assess the method performances on this kind of series correctly. Our search returned its use only for rainfall-runoff (in this case, the input is the rainfall, and the output is the runoff, so we cannot make comparisons). for rainfall–runoff (in this case, the input is the rainfall and the output is the runoff, so we cannot make comparisons). Moreover, other studies on modeling the Buzău River water flow have not been undertaken. Given the study’s novelty from this viewpoint, there are only a few articles on this topic that can be used for comparison [55,56,57].
It is important to indicate that a direct comparison of the goodness-of-fit parameters in this study with those in other papers analyzing different data series is not valid. This is because the algorithms must be executed in the same scenarios, including the same data series and training and test sets for AI models. Therefore, it is necessary to indicate that all the AI and ARIMA models, presented in extenso in [56,57], and compared below adhered to these criteria.
Comparison of the run time of the algorithms on S, S1, and S2 is given in Table 5, and the goodness-of-fit indicators are contained in Table 6.
The time taken to run SSA-ESN (and ESN) was the second (third) smallest, after that for ELM. The time to run MLP was between 2.11 s (for S2) and 5.11 s (for S).
Among the hybrid methods involving SSA, SSA-BP (SSA-ESN) performed best on the training sets of S and S1 (S2) in terms of MAE (MSE). With respect to MAE and MSE, SSA-BP outperformed SSA-ESN on the test sets of S and S1. When considering all goodness-of-fit indicators, SSA-ESN was the best on S2. Among all hybrid methods, CNN-LSM obtained remarkable results on the test set in terms of MAE and MSE. Taking R2 as the main criterion, BPNN was the worst model, followed by ELM. However, ELM best fit the S1 and S2 test sets with respect to MAE and S2 in terms of MAE and MSE. ESN and SSA-ESN were among the best models considering all criteria together. MLP and ARIMA gave the worst results among all methods.
The study’s main limitations are the following.
(1) The models do not capture the maxima.
According to [85], a complex method could be applied to address the first issue. The extreme events are memorized using a memory network, and the tail values are modeled by extreme value loss (EVL) [86]. Lastly, the results are combined to predict the raw data series.
(2) A significant increase in the time necessary to run the hybrid algorithm can be seen in Table 4 compared to ESN (Table 3). The same behavior was noticed when running SSA-ESN on daily data series (experiments will be presented in detail in another article): the ESN run time was 82.84 s for S, 6.15 s for S1, and 4.99 s for S2, and the SSA-ESN run time was 6505.60 s for S, 3805.99 for S1, and 3194.93 s for S2. Although SSA-ESN showed better performance in terms of MSE and R2, the increase in run time may not justify the marginal improvements in accuracy unless the task requires high precision.
This issue should be addressed by optimizing the parameters in the SSA algorithm.
Two other important topics would be of interest:
(a) Data preprocessing to remove inaccuracies (when data do not come from sources that have already performed the preprocessing, as in our case). This can be achieved using techniques involving decomposing and recomposing the data series [64,65,66,67].
(b) Replacement of the missing values (which was not the case here) by linear interpolation, average or median values of the recorded series, or use of an autoencoders.

4. Conclusions

This was a comprehensive investigation into the capability of two AI algorithms (ESN and SSA-ESN) to provide reliable models for the monthly river discharge of the Buzău River in Romania over 55 years. The results fill a gap in modeling hydro-meteorological data series using ESN [86,87]. More research should be done to confirm the suitability of ESN and SSA-ESN in this research field.
It has been proved that both models are comparable with respect to the goodness-of-fit indicators. The MAE values for both ESN and SSA-ESN were between 5.74 (for S2) and 7.60 (for S1) on the training set. On the test set, MAEs were in the range 4.48 (for S2) to 5.56 (for S1) on the test set. All R2 values were above 98.91%. MSE belongs to the interval [57.34, 102.94] on the training sets and to [36.60, 48.73] on the test sets, revealing that the algorithm performs better on the latter.
The removal of aberrant values considerably increases the algorithms’ run time from under 1 s to over 1 min (between 62.99 s and 155.59 s). The MSE values notably decreased on the training sets of all models, from the interval [57.34, 102.96] to the interval [39.73, 44.49]. Since the models did not fit the maxima well, future research will focus on better fitting for extremes and adapting the algorithm to reduce the run time.

Author Contributions

Conceptualization, A.B. and L.Z.; methodology, A.B. and L.Z.; software, L.Z.; validation, A.B.; formal analysis, A.B. and L.Z.; investigation, A.B. and L.Z.; resources, A.B.; data curation, A.B. and L.Z.; writing—original draft preparation, A.B.; writing—review and editing, A.B.; visualization, L.Z.; supervision, A.B.; project administration, A.B.; funding acquisition, A.B. All authors have read and agreed to the published version of the manuscript.

Funding

The research received no funding.

Data Availability Statement

Data will be available on request from the authors.

Conflicts of Interest

The present research carries no conflicts of interest.

References

  1. Gharehchopogh, F.S.; Namazi, M.; Ebrahimi, L.; Abdollahzadeh, B. Advances in Sparrow Search Algorithm: A Comprehensive Survey. Arch. Computat. Methods Eng. 2023, 30, 427–455. [Google Scholar] [CrossRef]
  2. Isfan, M.C.; Caramete, L.-I.; Caramete, A.; Basceanu, V.-A.; Popescu, T. Data analysis for gravitational waves using neural networks on quantum computers. Rom. Rep. Phys. 2023, 75, 113. [Google Scholar]
  3. Dai, Z.; Zhang, M.; Nedjah, N.; Xu, D.; Ye, F. A Hydrological Data Prediction Model Based on LSTM with Attention Mechanism. Water 2023, 15, 670. [Google Scholar] [CrossRef]
  4. Li, S.; Yang, J. Modelling of suspended sediment load by Bayesian optimized machine learning methods with seasonal adjustment. Eng. Appl. Comput. Fluid Mech. 2022, 16, 1883–1901. [Google Scholar] [CrossRef]
  5. Hayder, G.; Solihin, M.I.; Mustafa, H.M. Modelling of River Flow Using Particle Swarm Optimized Cascade-Forward Neural Networks: A Case Study of Kelantan River in Malaysia. Appl. Sci. 2020, 10, 8670. [Google Scholar] [CrossRef]
  6. Khan, M.Y.A.; Hasan, F.; Panwar, S.; Chakrapani, G.J. Neural network model for discharge and water-level prediction for Ramganga River catchment of Ganga Basin, India. Hydrol. Sci. J. 2016, 61, 2084–2095. [Google Scholar] [CrossRef]
  7. Samadi, M.; Sarkardeh, H.; Jabbari, E. Prediction of the dynamic pressure distribution in hydraulic structures using soft computing methods. Soft Comput. 2021, 25, 3873–3888. [Google Scholar] [CrossRef]
  8. Haghiabi, A.H.; Parsaie, A.; Ememgholizadeh, S. Prediction of discharge coefficient of triangular labyrinth weirs using adaptive neuro fuzzy inference system. Alex. Eng. J. 2018, 57, 1773–1782. [Google Scholar] [CrossRef]
  9. Bărbulescu, A.; Dumitriu, C.S. About the long-range dependence of cavitation effect on a copper alloy. Rom. J. Phys. 2024, 69, 904. [Google Scholar]
  10. Bărbulescu, A.; Dumitriu, C.S. Modeling the Voltage Produced by Ultrasound in Seawater by Stochastic and Artificial Intelligence Methods. Sensors 2022, 22, 1089. [Google Scholar] [CrossRef]
  11. Dumitriu, C.S.; Bărbulescu, A. Artificial intelligence models for the mass loss of copper-based alloys under the cavitation. Materials 2022, 15, 6695. [Google Scholar] [CrossRef] [PubMed]
  12. Dumitriu, C.Ş.; Dragomir, F.-L. Modeling the Signals Collected in Cavitation Field by Stochastic and Artificial Intelligence Methods. In Proceedings of the 2021 13th International Conference on Electronics, Computers and Artificial Intelligence (ECAI), Pitesti, Romania, 1–3 July 2021; pp. 1–4. [Google Scholar] [CrossRef]
  13. Adnan, R.M.; Jaafari, A.; Mohanavelu, A.; Kisi, O.; Elbeltagi, A. Novel Ensemble Forecasting of Streamflow Using Locally Weighted Learning Algorithm. Sustainability 2021, 13, 5877. [Google Scholar] [CrossRef]
  14. Van Thieu, N.; Nguyen, N.H.; Sherif, M.; El-Shafie, A.; Ahmed, A.N. Integrated metaheuristic algorithms with extreme learning machine models for river streamflow prediction. Sci. Rep. 2024, 14, 13597. [Google Scholar] [CrossRef]
  15. Alquraish, M.M.; Khadr, M. Remote-Sensing-Based Streamflow Forecasting Using Artificial Neural Network and Support Vector Machine Models. Remote Sens. 2021, 13, 4147. [Google Scholar] [CrossRef]
  16. Crăciun, A.; Costache, R.; Bărbulescu, A.; Chandra Pal, S.; Costache, I.; Dumitriu, C.S. Modern techniques for flood susceptibility estimation across the Deltaic Region (Danube Delta) from the Black Sea’s Romanian Sector. J. Marine Sci. Eng. 2022, 10, 1149. [Google Scholar] [CrossRef]
  17. Ferreira, R.G.; da Silva, D.D.; Elesbon, A.A.A.; Fernandes-Filho, E.I.; Veloso, G.V.; Fraga, M.D.S.; Ferreira, L.B. Machine learning models for streamflow regionalization in a tropical watershed. J. Environ. Manag. 2021, 280, 111713. [Google Scholar] [CrossRef]
  18. Piazzi, G.; Thirel, G.; Perrin, C.; Delaigue, O. Sequential Data Assimilation for Streamflow Forecasting: Assessing the Sensitivity to Uncertainties and Updated Variables of a Conceptual Hydrological Model at Basin Scale. Water Resour. Res. 2021, 57, 57. [Google Scholar] [CrossRef]
  19. Popescu, C.; Bărbulescu, A. On the Flash Flood Susceptibility and Accessibility in the Vărbilău Catchment (Romania). Rom. J. Phys. 2022, 67, 811. [Google Scholar]
  20. Saraiva, S.V.; Carvalho, F.D.O.; Santos, C.A.G.; Barreto, L.C.; Freire, P.K.D.M.M. Daily streamflow forecasting in Sobradinho Reservoir using machine learning models coupled with wavelet transform and bootstrapping. Appl. Soft Comput. 2021, 102, 107081. [Google Scholar] [CrossRef]
  21. Tanty, R.; Desmukh, T.S. Application of Artificial Neural Network in Hydrology—A Review. Int. J. Eng. Res. Techn. (IJERT) 2015, 4, 184–188. [Google Scholar]
  22. Tyralis, H.; Papacharalampous, G.; Langousis, A. Super ensemble learning for daily streamflow forecasting: Large-scale demonstration and comparison with multiple machine learning algorithms. Neural Comput. Appl. 2021, 33, 3053–3068. [Google Scholar] [CrossRef]
  23. Uca; Toriman, E.; Jaafar, O.; Maru, R.; Arfan, A.; Ahmar, A.S. Daily Suspended Sediment Discharge Prediction Using Multiple Linear Regression and Artificial Neural Network. J. Phys. Conf. Ser. 2018, 954, 012030. [Google Scholar] [CrossRef]
  24. Jaeger, H. Echo State Network. Available online: http://www.scholarpedia.org/article/Echo_state_network#Variants (accessed on 11 July 2024).
  25. Lark. Available online: https://www.larksuite.com/en_us/topics/ai-glossary/echo-state-network (accessed on 11 July 2024).
  26. Lukoševičius, M. A Practical Guide to Applying Echo State Networks. In Neural Networks: Tricks of the Trade. Lecture Notes in Computer Science; Montavon, G., Orr, G.B., Müller, K.R., Eds.; Springer: Berlin/Heidelberg, Germany, 2012; Volume 7700, pp. 659–686. [Google Scholar]
  27. Ma, Q.; Zhuang, W.; Shen, L.; Cottrell, G.W. Time series classification with Echo Memory Networks. Neural Netw. 2019, 117, 225–239. [Google Scholar] [CrossRef]
  28. Sohan, S.; Ozturk, M.C.; Principe, J.C. Signal Processing with Echo State Networks in the Complex Domain. In Proceedings of the 2007 IEEE Workshop on Machine Learning for Signal Processing, Thessaloniki, Greece, 27–29 August 2007; pp. 408–412. [Google Scholar]
  29. Skowronski, M.D.; Harris, J.G. Noise-Robust Automatic Speech Recognition Using a Predictive Echo State Network. IEEE Trans. Audio Speech Lang. Process. 2007, 15, 1724–1730. [Google Scholar] [CrossRef]
  30. Ibrahim, H.; Loo, C.K.; Alnajjar, F. Bidirectional parallel echo state network for speech emotion recognition. Neural Comput. Appl. 2022, 34, 17581–17599. [Google Scholar] [CrossRef]
  31. Daneshfar, F.; Jamshidi, M.B. A Pattern Recognition Framework for Signal Processing in Metaverse. In Proceedings of the 2022 8th Iranian Conference on Signal Processing and Intelligent Systems (ICSPIS), Behshahr, Iran, 28–29 December 2022; pp. 1–5. [Google Scholar] [CrossRef]
  32. Verstraeten, D.; Schrauwen, B.; D’Haene, M.; Stroobandt, D.D. An experimental unification of reservoir computing methods. Neural Netw. 2007, 2093, 391–403. [Google Scholar] [CrossRef]
  33. Li, Z.; Zheng, Z.; Outbib, R. Adaptive prognostic of fuel cells by implementing ensemble echo state networks in time-varying model space. IEEE Trans. Ind. Electron. 2019, 67, 379–389. [Google Scholar] [CrossRef]
  34. Morando, S.; Jemei, S.; Hissel, D.; Gouriveau, R.; Zerhouni, N. ANOVA method applied to proton exchange membrane fuel cell ageing forecasting using an echo network. Math. Comput. Simul. 2017, 131, 283–294. [Google Scholar] [CrossRef]
  35. Mezzi, R.; Yousfi-Steiner, N.; Péra, M.C.; Hissel, D.; Larger, L. An echo state network for fuel cell lifetime prediction under a dynamic micro-cogeneration load profile. Appl. Energ. 2021, 283, 116–297. [Google Scholar] [CrossRef]
  36. Jin, J.; Chen, Y.; Xie, C. Remaining useful life prediction of PEMFC based on cycle reservoir with jump model. Int. J. Hydrogen Energy 2021, 46, 40001–40013. [Google Scholar] [CrossRef]
  37. Morando, S.; Jemei, S.; Hissel, D.; Gouriveau, R.; Zerhouni, N. Show more Proton exchange membrane fuel cell ageing forecasting algorithm based on Echo State Network. Int. J. Hydrogen Energy 2017, 42, 1472–1480. [Google Scholar] [CrossRef]
  38. Jaeger, H.; Haas, H. Harnessing nonlinearity: Predicting chaotic systems and saving energy in wireless communication. Science 2004, 304, 78–80. [Google Scholar] [CrossRef] [PubMed]
  39. Ouyang, C.; Tang, F.; Zhu, D.; Qiu, Y.; Liu, Y. Application of improved sparrow search algorithm in concrete. J. Phys. Conf. Ser. 2021, 2082, 012014. [Google Scholar] [CrossRef]
  40. Fathy, A.; Alanazi, T.; Rezk, H.; Yousri, D. Optimal energy management of micro-grid using sparrow search algorithm. Energy Rep. 2022, 8, 758–773. [Google Scholar] [CrossRef]
  41. Wu, Z.; Wang, B. An ensemble neural network based on variational mode decomposition and an improved sparrow search algorithm for wind and solar power forecasting. IEEE Access 2021, 9, 166709–166719. [Google Scholar] [CrossRef]
  42. Song, J.; Jin, L.; Xie, Y.; Wei, C. Optimized XGBoost based sparrow search algorithm for short-term load forecasting. In Proceedings of the 2021 IEEE International Conference on Computer Science, Artificial Intelligence and Electronic Engineering (CSAIEE), Virtual Event, 20–22 August 2021; IEEE: New York, NY, USA, 2021; pp. 213–217. [Google Scholar]
  43. Lv, J.; Sun, W.; Wang, H.; Zhang, F. Coordinated approach fusing RCMDE and sparrow search algorithm-based SVM for fault diagnosis of rolling bearings. Sensors 2021, 21, 5297. [Google Scholar] [CrossRef]
  44. Xiong, Q.; Zhang, X.; He, S.; Shen, J. A fractional-order chaotic sparrow search algorithm for enhancement of long distance IRIS image. Mathematics 2021, 9, 2790. [Google Scholar] [CrossRef]
  45. Li, L.L.; Xiong, J.L.; Tseng, M.L.; Yan, Z.; Lim, M. Using multi-objective sparrow search algorithm to establish active distribution network dynamic reconfiguration integrated optimization. Expert Syst. Appl. 2022, 193, 116445. [Google Scholar] [CrossRef]
  46. Thenmozhi, R.; Nasir, A.; Sonthi, V.; Avudaiappan, T.; Kadry, S.; Pin, K.; Nam, Y. An improved sparrow search algorithm for node localization in WSN. Comput. Mater. Contin. 2022, 71, 2037–2051. [Google Scholar]
  47. Jiang, F.; Han, X.; Zhang, W.; Chen, G. Atmospheric PM2.5 prediction using Deepar optimized by sparrow search algorithm with opposition-based and fitness-based learning. Atmosphere 2021, 12, 894. [Google Scholar] [CrossRef]
  48. An, G.; Jiang, Z.; Chen, L.; Cao, X.; Li, Z.; Zhao, Y.; Sun, H. Ultra short-term wind power forecasting based on sparrow search algorithm optimization deep extreme learning machine. Sustainability 2021, 13, 10453. [Google Scholar] [CrossRef]
  49. Awadallah, M.A.; Al-Betar, M.A.; Doush, I.A.; Makhadmeh, S.N.; Al-Naymat, G. Recent Versions and Applications of Sparrow Search Algorithm. Arch. Computat. Methods Eng. 2023, 30, 2831–2858. [Google Scholar] [CrossRef]
  50. Mocanu-Vargancsik, C.A.; Bărbulescu, A. Analysis on Variability of Buzau River Monthly Discharges. Ovidius Univ. Ann. Ser. Civil Eng. 2018, 20, 51–55. [Google Scholar] [CrossRef]
  51. Mocanu-Vargancsik, C.A.; Bărbulescu, A. On the variability of a river water flow, under seasonal conditions. Case study. IOP Conf. Ser. Earth Environ. Sci. 2019, 344, 012028. [Google Scholar] [CrossRef]
  52. Mocanu-Vargancsik, C.; Tudor, G. On the linear trends of a water discharge data under temporal variation. Case study: The upper sector of the Buzău river (Romania). Forum Geogr. 2020, XIX, 37–44. [Google Scholar] [CrossRef]
  53. Minea, G.; Bărbulescu, A. Statistical assessing of hydrological alteration of Buzău River induced by Siriu dam (Romania). Forum Geogr. 2014, 13, 50–58. [Google Scholar] [CrossRef]
  54. Bărbulescu, Statistical Assessment and Model for a River Flow under Variable Conditions. Available online: https://cest2017.gnest.org/sites/default/files/presentation_file_list/cest2017_00715_poster_paper.pdf (accessed on 7 September 2024).
  55. Bărbulescu, A.; Mohammed, N. Study of the river discharge alteration. Water 2024, 16, 808. [Google Scholar] [CrossRef]
  56. Bărbulescu, A.; Zhen, L. Forecasting the River Water Discharge by Artificial Intelligence Methods. Water 2024, 16, 1248. [Google Scholar] [CrossRef]
  57. Zhen, L.; Bărbulescu, A. Comparative Analysis of Convolutional Neural Network-Long Short-Term Memory, Sparrow Search Algorithm-Backpropagation Neural Network, and Particle Swarm Optimization-Extreme Learning Machine for the Water Discharge of the Buzău River, Romania. Water 2024, 16, 289. [Google Scholar] [CrossRef]
  58. Chendeş, V. Water Resources in Curvature Subcarpathians. Geospatial Assessments; Editura Academiei Române: Bucureşti, Romania, 2011; (In Romanian with English Abstract). [Google Scholar]
  59. Updated Management Plan of the Buzau-Ialomita Hydrographic Area. Available online: http://buzau-ialomita.rowater.ro/wp-content/uploads/2021/02/PMB_ABABI_Text_actualizat.pdf (accessed on 17 October 2023). (In Romanian).
  60. Chai, Y.; Zhu, B.; Yue, Y.; Yang, Y.; Li, S.; Ren, J.; Xiong, H.; Cui, X.; Yan, X.; Li, Y. Reasons for the homogenization of the seasonal discharges in the Yangtze River. Hydrol. Res. 2020, 51, 470–483. [Google Scholar] [CrossRef]
  61. Chai, Y.; Yue, Y.; Zhang, L.; Miao, C.; Borthwick, A.G.L.; Zhu, B.; Li, Y.; Dolman, A.J. Homogenization and polarization of the seasonal water discharge of global rivers in response to climatic and anthropogenic effects. Sci. Total Environ. 2020, 709, 136062. [Google Scholar] [CrossRef] [PubMed]
  62. McManamay, R.A.; Orth, D.J.; Dolloff, C.A. Revisiting the homogenization of dammed rivers in the southeastern US. J. Hydrol. 2012, 424–425, 217–237. [Google Scholar] [CrossRef]
  63. Poff, N.L.; Olden, J.D.; Merritt, D.M.; Pepin, D.M. Homogenization of regional river dynamics by dams and global biodiversity implications. Proc. Nat. Acad. Sci. USA 2007, 104, 5732–5737. [Google Scholar] [CrossRef] [PubMed]
  64. Golyandina, N.; Korobeynikov, A. Basic Singular Spectrum Analysis and forecasting with R. Comp. Stat. Data An. 2014, 71, 934–954. [Google Scholar] [CrossRef]
  65. Golyandina, N.; Korobeynikov, A.; Zhigljavsky, A. Singular Spectrum Analysis with R; Springer: Berlin/Heidelberg, Germany, 2018. [Google Scholar]
  66. Nguyen, A.D.; Le Nguyen, P.; Vu, V.H.; Pham, Q.V.; Nguyen, V.H.; Nguyen, M.H. Accurate discharge and water level forecasting using ensemble learning with genetic algorithm and singular spectrum analysis-based denoising. Sci. Rep. 2022, 12, 19870. [Google Scholar] [CrossRef]
  67. Flandrin, P.; Rilling, G.; Goncalves, P. Empirical mode decomposition as a filter bank. IEEE Signal Proc. Lett. 2004, 11, 112–114. [Google Scholar] [CrossRef]
  68. Huang, N.E.; Shen, Z.; Long, S.R. The empirical mode decomposition and the Hilbert spectrum for nonlinear and non-stationary time series analysis. Proc. R. Soc. Lond. 1998, 454, 903–995. [Google Scholar] [CrossRef]
  69. Chu, T.-Y.; Huang, W.-C. Application of Empirical Mode Decomposition Method to Synthesize Flow Data: A Case Study of Hushan Reservoir in Taiwan. Water 2020, 12, 927. [Google Scholar] [CrossRef]
  70. Liu, F.; Li, J.; Liu, L.; Huang, L.; Fang, G. Application of the EEMD method for distinction and suppression of motion-induced noise in grounded electrical source airborne TEM system. J. Appl. Geophys. 2017, 139, 109–116. [Google Scholar] [CrossRef]
  71. Kwiatkowski, D.; Phillips, P.C.B.; Schmidt, P.; Shin, Y. Testing the null hypothesis of stationarity against the alternative of a unit root: How sure are we that economic time series have a unit root? J. Econ. 1992, 54, 159–178. [Google Scholar]
  72. Conover, W.J.; Johnson, M.E.; Johnson, M.M. A comparative study of tests for homogeneity of variances, with applications to the outer continental shelf bidding data. Technometrics 1981, 23, 351–361. [Google Scholar] [CrossRef]
  73. Pettitt, A.N. A non-parametric approach to the change point problem. J. Royal Stat. Soc. Ser. C Appl. Stat. 1979, 28, 126–135. [Google Scholar] [CrossRef]
  74. Fox, A.J. Outliers in Time Series. J. Royal Stat. Soc. Ser. B 1972, 34, 350–363. [Google Scholar] [CrossRef]
  75. Blázquez-García, A.; Conde, A.; Mori, U.; Lozano, J.A. A Review on Outlier/Anomaly Detection in Time Series Data”. ACM Comput. Surv. 2021, 54, 1–33. [Google Scholar] [CrossRef]
  76. AnomalyDetection R Package. Available online: https://github.com/twitter/AnomalyDetection/tree/master (accessed on 17 July 2024).
  77. Rosner, B. Percentage Points for a Generalized ESD Many-Outlier Procedure. Technometrics 1983, 25, 165–172. [Google Scholar] [CrossRef]
  78. Hochenbaum, J.; Vallis, O.S.; Kejariwal, A. Automatic Anomaly Detection in the Cloud Via Statistical Learning. 2017. Available online: https://arxiv.org/pdf/1704.07706 (accessed on 17 July 2024).
  79. Cleveland, W.S. Robust locally weighted regression and smoothing scatterplots. J. Am. Stat. Assoc. 1979, 74, 829–836. [Google Scholar] [CrossRef]
  80. Cleveland, R.B.; Cleveland, W.S.; McRae, J.E.; Terpenning, I.J. STL: A seasonal-trend decomposition procedure based on loess. J. Official Stat. 1990, 6, 3–33. [Google Scholar]
  81. He, K.; Mao, L.; Yu, J.; Huang, W.; He, Q.; Jackson, L. Long-term performance prediction of PEMFC based on LASSO-ESN. IEEE Trans. Instrum. Meas. 2021, 70, 3511611. [Google Scholar] [CrossRef]
  82. Jin, C.; Jiashu, J.; Yuepeng, C.; Changjun, X.; Bojun, L. PEMFC Performance Degradation Prediction Based on Bayesian Optimized ESN (In Chinese). Available online: https://www.researchgate.net/publication/380696856_jiyubeiyesiyouhuaESNdePEMFCxingnengtuihuayuce#fullTextFileContent (accessed on 11 July 2024).
  83. Sun, C.; Song, M.; Hong, S.; Li, H. A Review of Designs and Applications of Echo State Networks. Available online: https://arxiv.org/pdf/2012.02974 (accessed on 11 July 2024).
  84. Xue, J.; Shen, B. A novel swarm intelligence optimization approach: Sparrow search algorithm. Syst. Sci. Control Eng. 2020, 8, 22–34. [Google Scholar] [CrossRef]
  85. Ding, D.; Zhang, M.; Pan, X.; Yang, M.; He, X. Modeling Extreme Events in Time Series Prediction. Available online: http://staff.ustc.edu.cn/~hexn/papers/kdd19-timeseries.pdf (accessed on 8 September 2024).
  86. Zhang, M.; Ding, D.; Pan, X.; Yang, M. Enhancing Time Series Predictors With Generalized Extreme Value Loss. IEEE Trans. Knowl. Data Eng. 2023, 35, 1473–1487. [Google Scholar] [CrossRef]
  87. Yen, M.H.; Liu, D.W.; Hsin, Y.C.; Lin, C.E.; Chen, C.C. Application of the deep learning for the prediction of rainfall in Southern Taiwan. Sci. Rep. 2019, 9, 12774. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Map of the Buzău River catchment [53].
Figure 1. Map of the Buzău River catchment [53].
Water 16 02916 g001
Figure 2. The series S1 and S2, and the values of their basic statistics.
Figure 2. The series S1 and S2, and the values of their basic statistics.
Water 16 02916 g002
Figure 3. ESN’s structure.
Figure 3. ESN’s structure.
Water 16 02916 g003
Figure 4. (a) Data series (red) and its anomalies (blue dots). (b) The series without anomalies.
Figure 4. (a) Data series (red) and its anomalies (blue dots). (b) The series without anomalies.
Water 16 02916 g004
Figure 5. (a) ESN model on the test set for (a) S, (b) S1, and (c) S2.
Figure 5. (a) ESN model on the test set for (a) S, (b) S1, and (c) S2.
Water 16 02916 g005
Figure 6. (a) ESN model on the test set without aberrant values for (a) S, (b) S1, and (c) S2.
Figure 6. (a) ESN model on the test set without aberrant values for (a) S, (b) S1, and (c) S2.
Water 16 02916 g006
Table 1. Performance of the ESN models.
Table 1. Performance of the ESN models.
Series Training SetTest Set
Run Time (s)MAEMSER2 (%)MAEMSER2 (%)
S0.999276.6980.5899.765.0042.7199.69
S10.890327.60102.9598.915.5648.7499.16
S20.869295.7457.3599.484.4836.6199.52
Table 2. Performance of the SSA-ESN models.
Table 2. Performance of the SSA-ESN models.
Series Training SetTest Set
Run Time (s)MAEMSER2 (%)MAEMSER2 (%)
S0.966.6980.5899.755.0042.7299.68
S10.867.60102.9498.935.5648.7399.17
S20.795.7457.3499.484.4836.6099.53
Table 3. Performance of the ESN models for the series without aberrant values.
Table 3. Performance of the ESN models for the series without aberrant values.
Series Training SetTest Set
Run Time (s)MAEMSER2 (%)MAEMSER2 (%)
S_a2.275.1040.3399.684.4033.2899.72
S1_a1.066.0654.8794.694.8737.9596.34
S2_a0.944.9839.3399.994.4732.2199.99
Table 4. Performance of the SSA-ESN models for the series without aberrant values.
Table 4. Performance of the SSA-ESN models for the series without aberrant values.
Series Training SetTest Set
Run Time (s)MAEMSER2 (%)MAEMSER2 (%)
S_a155.594.8644.4991.914.4237.4191.32
S1_a62.995.4854.6492.844.6937.9793.83
S2_a110.314.4739.7390.954.1635.4690.59
Table 5. Comparisons of the run time on S, S1, and S2 [56,57].
Table 5. Comparisons of the run time on S, S1, and S2 [56,57].
MLPBPNNELMESNLSTMCNN-LSTMPSO-ELMSSA-BPSSA-ESN
S5.111.320.701.004.3310.1884.35475.430.96
S13.871.230.750.893.576.3557.37399.830.86
S22.111.160.650.873.595.8651.52435.160.79
Table 6. Comparisons of AI methods’ goodness-of fit indicators [56,57] on S, S1, and S2.
Table 6. Comparisons of AI methods’ goodness-of fit indicators [56,57] on S, S1, and S2.
MethodSeriesTraining SetTest Set
MAEMSER2 (%)MAEMSER2 (%)
S12.04221.1734.1310.90234.7926.80
ARIMAS113.92239.5645.2311.49260.6425.01
S210.22192.1834.1812.36299.0948.82
S10.16206.0535.769.75146.239.69
MLPS111.41252.1927.448.85135.7316.18
S29.26181.7036.9310.10158.14 2.33
ESNS6.6980.5899.765.0042.7299.69
S17.60102.9598.915.5648.7499.16
S25.7457.3599.484.4836.6199.52
ELMS6.0198.1283.054.6041.2988.70
S16.79126.3376.145.2154.5481.84
S25.0378.6379.714.0132.2189.71
LSTMS6.7987.6999.394.9241.4899.83
S110.51213.2298.997.6498.7499.74
S25.7260.0799.924.4935.6599.97
BPNNS6.96152.4452.895.52125.0631.07
S111.00326.6218.307.94116.3640.80
S28.14145.3850.218.29158.5542.17
SSA-ESNS6.6980.5899.755.0042.7299.68
S17.60102.9498.935.5648.7399.17
S25.7457.3499.484.4836.6099.53
SSA-BPS5.7391.2683.974.2932.5092.97
S17.00105.4092.765.2044.6296.12
S27.71132.4553.118.09168.6019.76
CNN-LSTMS6.0393.8189.454.2436.0094.58
S16.52115.0988.394.4839.9894.26
S24.7462.0093.013.5229.8395.04
PSO-ELMS6.0198.1383.054.6041.2888.68
S16.78126.5575.965.1352.1883.35
S25.0470.7079.663.9930.96889.94
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhen, L.; Bărbulescu, A. Echo State Network and Sparrow Search: Echo State Network for Modeling the Monthly River Discharge of the Biggest River in Buzău County, Romania. Water 2024, 16, 2916. https://doi.org/10.3390/w16202916

AMA Style

Zhen L, Bărbulescu A. Echo State Network and Sparrow Search: Echo State Network for Modeling the Monthly River Discharge of the Biggest River in Buzău County, Romania. Water. 2024; 16(20):2916. https://doi.org/10.3390/w16202916

Chicago/Turabian Style

Zhen, Liu, and Alina Bărbulescu. 2024. "Echo State Network and Sparrow Search: Echo State Network for Modeling the Monthly River Discharge of the Biggest River in Buzău County, Romania" Water 16, no. 20: 2916. https://doi.org/10.3390/w16202916

APA Style

Zhen, L., & Bărbulescu, A. (2024). Echo State Network and Sparrow Search: Echo State Network for Modeling the Monthly River Discharge of the Biggest River in Buzău County, Romania. Water, 16(20), 2916. https://doi.org/10.3390/w16202916

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop