Next Article in Journal
Review of Film Cooling Techniques for Aerospace Vehicles
Next Article in Special Issue
A Mid-Term Scheduling Method for Cascade Hydropower Stations to Safeguard Against Continuous Extreme New Energy Fluctuations
Previous Article in Journal
Proactive Regulation for Hydrogen Supply Chains: Enhancing Logistics Frameworks in Australia
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hybrid Multi-Branch Attention–CNN–BiLSTM Forecast Model for Reservoir Capacities of Pumped Storage Hydropower Plant

1
School of Electric Power Engineering, South China University of Technology, Guangzhou 510641, China
2
Branch Company of Maintenance & Test, China Southern Power Grid Energy Storage Co., Ltd., Guangzhou 510620, China
3
Guangzhou Power Electrical Technology Co., Ltd., Guangzhou 510535, China
4
School of Automation Science and Engineering, South China University of Technology, Guangzhou 510641, China
*
Author to whom correspondence should be addressed.
Energies 2025, 18(12), 3057; https://doi.org/10.3390/en18123057
Submission received: 3 May 2025 / Revised: 27 May 2025 / Accepted: 3 June 2025 / Published: 10 June 2025
(This article belongs to the Special Issue Optimal Schedule of Hydropower and New Energy Power Systems)

Abstract

Pumped storage hydropower plants are important resources for scheduling urban energy storage, which realize the conversion of electric energy through upper and lower reservoir capacities. Dynamic forecasting of reservoir capacities is crucial for scheduling pumped storage and maximizing the economic benefits of pumped storage hydropower plants. In this work, a hybrid forecast network is proposed for both the upper and lower reservoir capacities of a pumped storage hydropower plant. A bidirectional long- and short-term memory network (BiLSTM) is designed as the baseline for the prediction model. A convolutional neural network (CNN) and Squeeze-and-Excitation (SE) attention mechanism are designed to extract local features from raw time series data to capture short-term dependencies. In order to better distinguish the effects of different data types on the reservoir capacity, the correlation between data and reservoir capacity is analyzed using the Spearman coefficient, and a multi-branch forecast model is established based on the correlation. A fusion module is designed to weight and fuse the branch prediction results to obtain the final reservoir capacities forecast model, namely, Multi-Branch Attention–CNN–BiLSTM. The experimental results show that the proposed model exhibits better forecast accuracy in forecasting the reservoir capacity compared with existing methods. Compared with BiLSTM, the M A P E of the forecast values of the reservoir capacities of the upper and lower reservoirs decreased by 1.93 % and 2.2484 % , the R M S E decreased by 16.9887 m 3 and 14.2903 m 3 , and the R 2 increased by 0.1278 and 0.1276 , respectively.

1. Introduction

Among existing large-scale energy storage technologies, pumped storage power plants are the most attractive large-scale energy storage solution, due to the vast potential energy contained in their reservoirs, the high full-cycle energy conversion efficiency, the low unit cost of generation, and the flexibility they provide to transmission system operators in short-term operations [1]. Pumped storage hydropower plants are important energy storage resources in the power system, which realize the recovery of electricity by mobilizing the water volume in reservoirs [2,3]. The main components of a pumped storage system include an upper reservoir (UR) and a lower reservoir (LR) for water storage, pressure pipes connecting the UR and LR, and pumps to lift water from the lower reservoir to the upper reservoir [4]. Reservoir capacity is a core parameter of pumped storage hydropower plant operation, in which the upper reservoir is responsible for storing potential energy and the lower reservoir is used to collect the returning water. The capacity determines the energy storage capacity and operation efficiency of the pumping station [5]. By using the electric power for pumping at the low point of electricity consumption, the water is stored in the upper reservoir, and releasing the water generates electricity at the peak point of electricity consumption. Pumped storage hydropower plants can provide the grid with functions such as peak shifting, valley filling, and emergency backup [6]. A schematic of a pumped storage hydropower plant is shown in Figure 1. The reservoir capacities are affected by many factors, including the released/saved water, weather, temperature, and soil and water permeability. Pumped storage accounts for approximately 96% of global stored electricity capacity and 99% of global stored energy [7]. An accurate forecast of upper and lower reservoir capacity changes is important for realizing efficient scheduling and safe pumping station operation. The accurate forecasting of reservoir capacity can not only optimize the scheduling strategy of a pumping station but also reasonably allocate the water scheduling of the upper and lower reservoirs, improving the overall energy efficiency of pumping station operation. Therefore, the combination of advanced data-driven forecast methods can comprehensively explore the complex nonlinear relationship between multiple variables and provide a scientific basis for decision makers, which effectively guarantees the efficient and sustainable operation of pumped storage hydropower plants.
Data correlation analysis is a statistical method used to assess relationships between variables and is widely used in scientific research and practical decision-making [2]. The correlation between input data and output data needs to be considered when making predictive models, and data with higher correlation can improve the predictive accuracy of the model. Common methods include the Pearson correlation coefficient [8], Spearman correlation coefficient [9], and Kendall correlation coefficient [10]. The Pearson correlation coefficient is used to measure the linear relationship between two continuous variables and is suitable for data that are normally distributed and linear [11]. The Spearman correlation coefficient and the Kendall correlation coefficient are based on the ordering of the data and are able to capture nonlinear but monotonic relationships and are suitable for the analysis of data that have a high number of outliers or are not normally distributed [12]. The development of prediction models has experienced a gradual evolution from traditional statistical methods to modern deep learning models [13]. Traditional mathematical models such as linear regression [14], autoregressive (AR) analysis [15], and the differential autoregressive moving average (ARMA) [16] are widely used because of their simplicity and efficiency. However, these methods have limited ability to portray nonlinear features and complex relationships [17]. Subsequently, Support Vector Regression (SVR) [18] and Random Forest (RF) [19], based on integrated learning and gradient-boosted decision trees [20], have emerged, enhancing the ability of modeling nonlinear relationships but performing slightly less well in capturing time dependencies.
With the development of neural networks, such as back-propagation neural networks (BPNNs) [21] and artificial neural networks (ANNs) [22], the ability of models to recognize complex nonlinear features is further improved [23]. In particular, deep learning models such as convolutional neural networks (CNNs) [24,25,26,27] and long short-term memory (LSTM) networks [28,29] have gradually become mainstream, with CNNs being able to efficiently extract local features and LSTM being effective at capturing long-term and short-term temporal dependencies. LSTM effectively solves the problem of gradient vanishing in traditional recurrent neural networks (RNNs) in long sequence learning through its gating mechanism, while bidirectional LSTM (BiLSTM) further develops this structure to capture richer contextual information in the time series by simultaneously processing the sequence data in both the forward and backward directions [30], thus improving the comprehensiveness and accuracy of prediction [31]. In order to obtain more accurate prediction models, hybrid models combining CNNs with LSTM or BiLSTM are widely used, which significantly improve the prediction accuracy by integrating the advantages of CNNs and LSTM [32]. The pattern of partial discharge for overhead covered conductors is detected by combing LSTM and GRUs together [33].
To further enhance the model’s ability to focus on key features, researchers introduced an attention mechanism to dynamically adjust the feature channel weights in a prediction model. The attention mechanism was implemented mainly by assigning weights to map the importance of different parts of the input to a weight vector [34]. The common methods for computing attention are mainly dot product attention and additive attention [35]. These methods obtain the weights associated with the input by performing appropriate transformations and similarity calculations on the input features. The Squeeze-and-Excitation (SE) attention mechanism is widely used in CNNs. In recent years, the multi-branch network structure has also received increasing focus [36]. This structure is able to extract the feature representation of each type of information in a targeted way by introducing multiple independent branch networks to specialize different input features.
In order to better manage the scheduling of upper and lower reservoir capacities in pumping stations, this paper establishes a Multi-Branch Attention–CNN–BiLSTM forecast model for predicting upper and lower reservoir capacities at future moments. Specifically, BiLSTM is used as the baseline forecast model, and a CNN is introduced to extract local features from raw time series data to capture short-term dependencies. An SE attention mechanism is used to adaptively adjust the weights of the forecast network. In order to further analyze the relationship between the data, the correlation between the input data and the output data is analyzed by using Spearman’s eigenfactors. A three-branch structure is set up and the branch forecast results are weighted to obtain the final forecast results, forming the Multi-Branch Attention–CNN–BiLSTM network. The main contributions of this paper can be summarized as follows:
(1)
An improved Multi-Branch Attention–CNN–BiLSTM network is proposed for reservoir capacity forecasting to improve feature extraction ability and enhance accuracy. A weighted fusion module is proposed to fuse the forecasting results. The proposed Multi-Branch Attention–CNN–BiLSTM framework can better capture nonlinear and non-stationary features of the data samples.
(2)
Spearman’s coefficient is presented to analyze the input–output data relationship to optimize feature selection and focus on the most relevant features. SE is designed to adaptively adjust the weights of the network, thereby enhancing the model’s ability to focus on the most relevant features for more accurate forecast.
(3)
A real application validation in a hydropower generation station shows that the proposed model can enhance model robustness and improve accuracy for reservoir capacity forecasting.
The outline of this work is given as follows. Data Descriptions and Pre-Processing is given in Section 2 with details of data descriptions and pre-processing. The Multi-Branch Attention–CNN–BiLSTM model is presented for reservoir capacity forecasting in Section 3. In Section 4, experimental results and analysis are given to show the effectiveness of the proposed method. Section 5 concludes this work.

2. Data Descriptions and Pre-Processing

2.1. Construction of Dataset

In this paper, real-time operational data from January 2023 to May 2023 of a pumped storage hydropower plant were used. The sample time interval is 30 min. The pumping station has an upper reservoir and a lower reservoir, and four speed-regulated pumps are installed in the pipeline connecting the upper and lower reservoirs to accomplish power generation and water transfer operations. The power and speed of each pump can be varied by regulation. The available data include the power and speed of each pump and the water level in the upper and lower reservoirs. Data were collected using sensors and transferred in real time to a database for storage and easy reading. The data types and units are shown in Table 1, where U 01 U 04 are the models of the four pumps, and the reservoir levels are represented by 1 and 2 for the upper and lower reservoirs, respectively. Pumps U 01 and U 03 are the connecting pumps of the pipeline from the lower reservoir to the upper reservoir for pumping water. Pumps U 02 and U 04 are connecting pumps for the pipeline from the lower reservoir to the upper reservoir for power generation. Both the upper and lower reservoir levels are the water level of the reservoir capacity. Reservoir levels 1 and 2 indicate the height of the reservoir level measured by the two sensors. Reservoir capacity indicates the volume of water in the reservoir. The forecasting of the capacities of the upper and lower reservoirs is investigated. The capacity distributions of the upper and lower reservoirs are shown in Figure 2. It can be seen that there is a certain periodicity in the distribution of the upper and lower reservoir capacities.

2.2. Data Pre-Processing

The dataset has some data with large deviations, which are due to errors caused by the sensors. In this paper, simple data processing was carried out on the data to eliminate the data which is beyond the normal range. Due to the excessive amount of data collected, not all of it can be used for reservoir forecasting. In this paper, the feature selection method of the Spearman coefficient was used to select the parameters which are more correlated with the reservoir capacities. The Spearman feature selection coefficient is a statistical measure of the relationship between variables. The Spearman coefficient assesses the correlation of variables by comparing their ranks. Specifically, the features and target variables are first ranked separately and their rank differences are calculated; then the value of the Spearman correlation coefficient is calculated based on these rank differences. The value of the coefficient ranges from [ 1 , 1 ] , where a positive value indicates a positive correlation between the variables and a negative value indicates a negative correlation. The closer to ± 1 it is, the stronger the correlation is. The specific expression of the Spearman feature selection coefficient is given as follows:
ρ = 1 6 d i 2 n n 2 1
where d i denotes the difference between the values of the ith data pair; n is the total number of samples. Assuming that X and Y are two variables, the characteristic coefficients between X and Y are to be analyzed. Then, d i = X i Y i , where X i is the ith datum of X; Y i is the ith datum of Y.
The various variables of the dataset, including the power and speed of each pump, the water level of the upper and lower reservoirs, and the reservoir capacity of the upper and lower reservoirs were subjected to Spearman characteristic coefficient selection for analyzing the correlation between these variables and the reservoir capacity of the upper and lower reservoirs. Since the output of the forecast model used in this paper includes both upper and lower reservoirs, the correlation coefficients of the variables are obtained separately for the upper and lower reservoirs. In order to obtain a common correlation coefficient, the two correlation coefficients are averaged after taking the absolute values. The Spearman correlation coefficients of each variable with the upper and lower reservoirs are shown in the Table 2.
A correlation coefficient between 0.8 and 1.0 indicates a very strong correlation between the data; between 0.6 and 0.8 indicates a strong correlation between the data; between 0.4 and 0.6 indicates a moderately strong correlation between the data; between 0.2 and 0.4 indicates a weak correlation between the data; and between 0.0 and 0.2 indicates a very weak correlation or no correlation between the data. In order to ensure as much as possible that the model learned the trends of the upper and lower reservoirs, we chose variables with average correlation coefficients greater than 0.3 as inputs to the forecast model, including the power of U 01 , the power of U 03 , the upper and lower reservoir levels, and the upper and lower reservoir capacities. As can be seen from the Table 2, the correlation between the reservoir capacity and itself and the water level is greater. For pumping units, the correlation is greater for pumping units U 01 and U 03 , due to the indirect relationship between pumping and water level; whereas pumping units that generate electricity affect the water level less.
In this paper, the five historical sets of sampled data were used to forecast the upper and lower reservoir capacity in the next 30 min. In order to meet the data format requirements of the time series forecast model, the five historical sets of sampled data were combined in time order with an input dimension of 40 and output dimension of 2. The data were divided into training data and test data according to 9:1. In order to meet the input requirements of the CNN, this paper performed data tiling on the one-dimensional data. If the number of data inputs was N, the number of one-dimensional data was 40 × N , which was tilted to 40 × 1 × 1 × N for the CNN. Normalization is a commonly used data preprocessing technique to transform data with different scales or value ranges into a uniform range to eliminate the impact of high-magnitude differences between features on model training and forecasting. In machine learning and deep learning tasks, the input data usually have different scales or distributions, and without normalization, variables with larger eigenvalues may dominate the optimization process of the objective function, leading to degradation of model performance or even failure to converge. Normalization can accelerate the gradient descent optimization process, improve the speed of model convergence, and increase the accuracy and robustness of specific algorithms. In this paper, all elements of the data are mapped between 0 and 1 to ensure scale consistency between data features. The normalization formula is as follows:
x = x x m i n x m a x x m i n
where x m i n and x m a x are the minimum and maximum values of the feature, respectively.

3. Multi-Branch Attention–CNN–BiLSTM Design for Upper and Lower Reservoir Capacity Forecasting

In this paper, a Multi-Branch Attention–CNN–BiLSTM network is designed for upper and lower reservoir capacity forecasting. The model takes BiLSTM as the baseline forecast model and introduces the CNN to improve the feature extraction ability of the model on the time series, which helps the model to better handle the depth information of the time series. To further improve the model’s ability to pay attention to data processing, the CNN is tuned by using the SE attention mechanism. The data used in this paper include reservoir capacities, water levels, and pumping machine power, and different data have different correlations for reservoir capacities.
In order to better extract the relationship between different data and reservoir capacity, this paper adopted a multi-branch network model, and each branch adopts the same Attention–CNN–BiLSTM structure. The branches were fed with different data types, respectively, and one of the branches included all data types. Each branch obtained the forecast values of the reservoir capacities, and finally all the forecast values were fused together by the weighted sum to obtain the final forecast results. The structural framework of the Mulit-Branch Attention–CNN–BiLSTM forecast model is shown in Figure 3.
A three-branch structure was designed with one input for pump power, one input for reservoir capacity and level, and one input for all data. The weighted sum of the fusion modules uses the following formula:
Y = a 1 Y 1 + a 2 Y 2 + a 3 Y 3
where Y 1 , Y 2 , and Y 3 are the forecast outputs of the first, second, and third branches, respectively; a 1 , a 2 , and a 3 are the weights of the first, second, and third branches, respectively; and a 1 + a 2 + a 3 = 1 . The corresponding weights were calculated based on the correlation coefficients in Table 2. The specific steps were calculating the average of the correlation coefficients of each channel and then calculating the weights based on the proportion of the three average correlation coefficients. The expression for weight calculation is
a i = s i s 1 + s 2 + s 3
where a i is the weight of the ith channel; s i is the average correlation coefficient of the ith channel; and s 1 , s 2 , s 3 are the average correlation coefficients of the three channels. In this paper, we used Spearman’s characteristic coefficient analysis and finally set a 1 = 0.2444 ; a 2 = 0.3969 ; and a 3 = 0.3587 .

3.1. Bidirectional Long- and Short-Term Memory Network

LSTM is a special RNN designed to solve the long-term dependency problem in sequence data. Although traditional RNNs are capable of handling sequence data, they perform poorly when dealing with long sequences due to the gradient vanishing or explosion problem. LSTM is able to alleviate this problem by introducing a gating mechanism which can effectively selectively remember or forget information. In addition, LSTM adds the cell state number C as a memory unit to correlate information over a longer time span. The core of LSTM consists of the forget gate, input gate, output gate, and cell state. The structural framework of the LSTM layer is shown in Figure 4.
(1)
Forget gate: The forget gate is used to control which historical information needs to be forgotten. A weight between 0 and 1 is generated for adjusting the forgetting ratio by weighting and summing the hidden state of the previous time step h t 1 and the current input x t as well as activating it. The expression for the equation of the forget gate is given:
f t = σ W f · h t 1 , x t + b f
where W f is the weight; b f is the bias; and σ is the Sigmoid function. The expression for the equation of the Sigmoid function is given as follows:
f ( x ) = 1 1 + e x
(2)
Input gate: The input gate determines which parts of the information from the current time step need to be added to the cell state and regulates the strength of the information update by an activation function that combines the current input and the previous hidden state. The expressions for the equations of the input gate are given as follows:
i t = σ W i · h t 1 , x t + b i
C ˜ t = tanh W C · h t 1 , x t + b C
where W i and W C are weights; b f and b C are biases; and tanh is the Tanh function. The expression for the equation of the Tanh function is
f ( x ) = e x e x e x + e x
(3)
Output gate: The output layer is used to regulate and output the hidden state output from the LSTM in the current time step, which contains extensive time-dependent information and is adapted to the current time step. The expression for the equations of the output gate are
o t = σ W o · h t 1 , x t + b o
h t = o t · tanh C t
where W o is the weight; b o is the bias; and C t is the current cell state number.
(4)
Cell state: The cell state is an important carrier of LSTM memory, which keeps the long-term memory of sequence data by discarding part of the old information through the forget gate and then introducing new information in combination with the input gate. The expression for the equation of the cell state is
C t = f t · C t 1 + i t · C ˜ t
The structure of LSTM can only process sequence data from front to back sequentially, which may ignore the effect of future time steps on current time steps. In many practical applications, however, future information often plays an important role in current judgment as well. For this reason, scholars have researched BiLSTM. BiLSTM processes sequence data from two directions by using two independent LSTM layers in both forward and backward directions. Forward LSTM extracts forward-dependent features of the time step from the sequence start point to the end point; backward LSTM extracts backward-dependent features of the time step from the sequence end point to the start point. The features in these two directions are eventually weighted and merged to generate a complete contextual representation of each time step, thus improving the model’s ability to understand the sequence data. The structural framework of BiLSTM is shown in Figure 4. Both forward and backward LSTMs learn the input data to obtain their respective hidden states with long-term dependence. BiLSTM connects the two hidden states, allowing the model to understand the time-series data in a more comprehensive way, which is suitable for complex forecast tasks.

3.2. Convolutional Neural Network and SE Attention Mechanism Design

In the regression forecast for reservoir capacities, the CNN can efficiently extract local patterns in time series data and reduce noise interference and provide high-quality feature representations for the subsequent forecast process. The CNN is mainly responsible for extracting local patterns and key features in the input data in CNN-BiLSTM-based regression forecasting, which provides more effective feature representation for subsequent time-dependent modeling. The CNN extracts local feature patterns from the input data through convolutional layers, such as short-term trends and key changes in the time series, and significantly reduces the computational complexity of the network by utilizing parameter sharing and local connectivity. The superposition of multiple convolutional layers enables the model to progressively extract more abstract features from low to high levels. In order to capture nonlinear relationships, the output after convolution is usually nonlinearly mapped by an activation function (e.g., ReLU), which provides the model with stronger expressive capability. To better extract the feature relationship between the data, this paper uses the connection of three convolutional layers, and the convolutional layers are connected by the ReLU activation function.
In traditional neural networks, the model processes the input data uniformly and does not differentiate between the attention paid to different parts. However, when processing longer time series, some parts may be more important than others. The purpose of the attention mechanism is to enable the model to selectively focus on the interested or critical parts of the data sequence, enabling further improvements in model accuracy. In this paper, we use the SE attention mechanism to improve the CNN network to help the model accurately capture key features and effectively reduce the interference of irrelevant information in the model. The SE attention mechanism is a lightweight but efficient channel attention mechanism that focuses on improving the feature representation of the CNN. Traditional CNNs assign the same importance to all channels by default when processing features, but in real tasks the contribution of different channels to the final goal often has differences.The SE attention mechanism improves the performance of the model by dynamically adjusting the weights of each channel to emphasize the important feature channels, while suppressing irrelevant or redundant channel information. The SE module consists of three parts: global information embedding (Squeeze), adaptive recalibration (Excitation), and reweighting (Scale). The structure of SE attention in the CNN is shown in Figure 5.
In the SE attention mechanism, the number of channels in the previous fully connected layer (FC) is 1/8 of the number of channels in the latter FC; the number of channels in the latter FC is equal to the number of channels in the last CNN layer. In the Squeeze stage, the SE module compresses the spatial information of each channel through global average pooling, generates the global feature description and extracts the key channel correlations; in the Excitation stage, the dependencies between channels are learned through a series of fully connected layers and activation functions, and the weighting coefficients are generated. In the Scale stage, the weights obtained earlier are reused on the original feature map to realize the dynamic adjustment of the channel features, thus improving the expressive ability of the model.

3.3. Multi-Branch Network for Attention–CNN–BiLSTM

To fully exploit the role of multi-source information in the forecast task, the multi-branch network structure has been widely adopted in SE Attention–CNN–BiLSTM forecast networks. By building a branch network to receive different categories of factors related to the target variable separately, a comprehensive input branch is introduced at the same time which is used to capture the global pattern of the overall features. The branch network structure is shown in Figure 6.
The structure of each branch of a multi-branch network is kept consistent by extracting feature representations of different inputs and fusing them in subsequent layers. Weighted fusion, an attention mechanism, or feature splicing are usually adopted in the fusion stage to ensure that the importance of different sources of information in decision making is reasonably expressed. Compared with the traditional single structure, models based on multi-branch design can more effectively separate and utilize relevant information from multi-source data, while demonstrating significant performance advantages in scenarios with complex data features or diverse sources. This structure provides a general and efficient solution for reservoir capacity forecasting.

4. Experiment Results

In this paper, Multi-Branch Attention–CNN–BiLSTM was used to predict the capacity of the upper and lower reservoirs. The input data included the power of U01, the power of U03, and the water level and capacity of the upper and lower reservoirs in the last five steps, each with a time interval of 30 min. The output data were the upper and lower reservoir capacities in the next step. The dimension of input data was 40 and the dimension of output data was 2. The training data were converted to 6473 rows and the test data were 341 rows according to the model data requirements. Multi-Branch Attention–CNN–BiLSTM was trained using the training data and the model was tested by using the test data. The model was validated by using metrics such as the M A P E , the R M S E , and R 2 , as well as scatter plots and distribution plots, to ensure correct forecasting of the model.

4.1. Model Training

In order to conduct the model training, this paper set appropriate Multi-Branch Attention–CNN–BiLSTM model parameters. The Multi-Branch Attention–CNN–BiLSTM model parameters are shown in Table 3. To balance the capability of the model and the computational complexity, 32 BiLSTM hidden-layer neurons were set in this paper. The number of channels of the CNN layer was set to 64 channel counts for the first layer and 32 for both the second and the third layers, which helped the model to extract multi-layered features from the original data and reduced the computational complexity and the risk of overfitting. To ensure the reliability of the weighted sums of the different branches, this paper scaled the division based on the correlation between the input data and the output data. The correlations of the input data of the three branches were averaged and the three averages were converted to a value whose sum was equal to one. The final weighted ratios were a 1 = 0.2444 ; a 2 = 0.3969 ; and a 3 = 0.3587 .

4.2. Model Testing

In this paper, the M A P E , the R M S E , and R 2 were used to statistically forecast the accuracy of the model. The M A P E is the mean absolute percentage error, which does not change due to global scaling of the variables. R M S E is the root mean square error, which measures the deviation of forecast values from actual observations. R 2 is the coefficient of determination, which is a visual representation of the accuracy of the model and can take values between 0 and 1. The smaller the values of the M A P E and R M S E are, the higher the forecast accuracy of the model is; the closer R 2 is to 1, the higher the forecast accuracy of the model is; and the closer R 2 is to 0, the worse the forecast accuracy is. The expressions of the M A P E , the R M S E , and R 2 are
M A P E = 100 % n i = 1 n y ^ i y i y i
R M S E = 1 n i = 1 n y ^ i y i 2
R 2 = 1 i = 1 n y ^ i y i 2 i = 1 n y ¯ i y i 2
where y ^ i is the forecast value; y i is the actual value; y ¯ i is the average value; and n is the size of the data.
To verify the effectiveness of Multi-Branch Attention–CNN–BiLSTM in this work, multiple models were used for comparison including the original LSTM model, the original BiLSTM model, CNN-improved BiLSTM (CNN-BiLSTM), and SE attention mechanism-improved CNN-BiLSTM (SE-CNN-BiLSTM). In addition, in order to verify that the LSTM family of models has a better forecast ability for time series data, the BPNN model was introduced for comparison. The forecasts of the upper reservoir, lower reservoir, and average capacities in different models are respectively shown in the Table 4, Table 5, and Table 6. The distribution and scatter plots of the forecast values of different models on the test data are plotted. The distribution and scatter plots of the upper reservoir capacity forecast results are shown in Figure 7 and Figure 8. The distribution and scatter plots of the lower reservoir capacity forecast results are shown in Figure 9 and Figure 10.
As can be seen from Figure 7, Figure 8, Figure 9 and Figure 10, the Multi-Branch SE–CNN–BiLSTM designed in this paper had the best results in the double-output forecast of the upper and lower reservoirs, and our model could complete the forecast of the upper and lower reservoir capacities. This is not only reflected in its ability to follow the trend of the actual data but also in the alignment of the predicted values with the actual values in the scatter distribution plots. In Figure 7 and Figure 9, the predicted curves of the Multi-Branch SE–CNN–BiLSTM model are closest to the actual data curves, indicating that the model was able to effectively capture the dynamic changes in the data. In Figure 8 and Figure 10, the point-to-point distributions of predicted and actual values are closest to the diagonal line, which further validates the high accuracy of the Multi-Branch SE–CNN–BiLSTM model model predictions. Both LSTM and BiLSTM achieved better forecast results compared to the BPNN, which indicates that the BPNN is not ideal for performing time series processing. From the tab4,tab5,tab6, it can be seen that Multi-Branch SE–CNN–BiLSTM predicted the upper reservoir capacity with an R 2 of 0.9231, M A P E of 2.8853%, and R M S E of 26.8869 m 3 ; and it predicted the lower reservoir capacity with an R 2 of 0.8242, M A P E of 5.4251%, and R M S E of 45.5623 m 3 . The average metrics for the two reservoir capacities were an R 2 of 0.8737, M A P E of 4.1552%, and R M S E of 36.2246 m 3 . The computing time of Multi-Branch Attention–CNN–BiLSTM was only 0.0525 s, which is very short time for the considered problem, although it required more computing time compared to other models. This is due to the fact that Multi-Branch Attention–CNN–BiLSTM has a more complex model architecture.
The introduction of the CNN and SE attention mechanisms substantially improves the forecast accuracy of the model. The upper and lower reservoir capacity M A P E of CNN-BiLSTM was improved by 0.9316% and 0.8452% compared to that of BiLSTM, respectively, and the mean of the two was improved by 0.8884%; the upper and lower reservoir capacity M A P E of SE-CNN-BiLSTM was improved by 1.0209% and 0.9396% compared to that of CNN-BiLSTM, and the mean of both was improved by 0.9802%. The introduction of the multi-branch model improved the forecast accuracy of the lower reservoir capacity and further ensured the accuracy of the double-output model forecast. The introduction of the multi-branch model made the model forecast values more closely match the distribution of the actual data, which indicates the reliability of multi-branch improvement in the double-output forecast of the upper and lower reservoirs.
The 3 σ rule is used to assess the reliability of the forecast of a predictive model. In statistics, data are normally distributed in principle, and the percentage of values that are less than one standard deviation, two standard deviations, and within three standard deviations from the mean are the precise figures of 68.27%, 95.45%, and 99.73%. Therefore, it can be assumed that a data value is almost distributed within the interval ( μ 3 σ , μ + 3 σ ) , and the probability of exceeding this range is less than 0.3%. After the forecast model generates results, the mean and standard deviation of the distribution of forecast values can be calculated as a way of determining whether certain forecast values are outside a reasonable range. When the forecast result exceeds three times the standard deviation, it indicates that there is an error in the model or an anomaly in the forecast target data. By analyzing these anomalous results, it can be found to provide a warning for the forecast target to avoid loopholes or equipment problems in the daily production process.
In this paper, the difference between the actual and forecast data was calculated separately for the upper and lower reservoir capacities, and the 3 σ ReLu historgram of the difference was analyzed. The mean value of the upper reservoir capacity residual was 11.9863 m 3 and the standard deviation was 24.1193 m 3 ; the mean value of the lower reservoir capacity residual was 7.6263 m 3 and the standard deviation was 45.0166 m 3 . The 3 σ ReLu historgram of the error of the upper and lower reservoir capacity is shown in Figure 11. Therefore, the error distribution of the upper reservoir capacity is ( 60.3716 m 3 ,   84.3442 m 3 ) , and the range of error distribution for the lower reservoir capacity is ( 127.4235 m 3 ,   142.6761 m 3 ) . When the residuals exceed these ranges, there may be a problem with the upper and lower reservoir capacities, such as leakage. This information allows workers to provide real-time maintenance to ensure the safe operation of a pumped storage hydropower plant.

5. Conclusions

In this paper, the Multi-Branch Attention–CNN–BiLSTM forecast model was developed to predict the upper and lower reservoir capacities of pumped storage hydropower plants. The correlation of the data was analyzed using Spearman’s characteristic coefficients, and the power of U 01 , the power of U 03 , and the level and the capacity of the upper and lower reservoirs were selected to predict the capacity of the upper reservoir. The experimental results show that the Multi-Branch Attention–CNN–BiLSTM model exhibits higher accuracy in predicting the reservoir capacity of pumped storage hydropower plants compared to the baseline model. Compared with the original BiLSTM, the forecast results were closest to the trend of the test data, with an increase of 1.9300 % and 2.2484 % in the M A P E , 16.9887 m 3 and 14.2903 m 3 in the R M S E , and 0.1278 and 0.1276 in R 2 , respectively. Using the Multi-Branch Attention–CNN–BiLSTM model, the average metrics for the two reservoir capacities were an R 2 of 0.8737, M A P E of 4.1552%, and R M S E of 36.2246 m 3 . The error distribution ranges of the upper and lower reservoir capacities were obtained based on 3 σ as ( 60.3716 m 3 , 84.3442 m 3 ) and ( 127.4235 m 3 , 142.6761 m 3 ) , respectively, and such information can provide a real-time maintenance strategy to ensure the safe operation of pumped storage hydropower plants. The prediction model proposed in this paper has dual outputs, and the change in the lower reservoir capacity is more unstable than the change in the upper reservoir capacity, which leads to a larger error range in the lower reservoir capacity.

Author Contributions

Conceptualization, Y.G. and Y.Z.; Methodology, Y.G. and J.Z.; Validation, Y.G.; Formal analysis, Y.G.; Investigation, L.Z.; Resources, H.W.; Data curation, Y.G. and H.W.; Writing—original draft, Y.G.; Writing—review & editing, H.W., J.Z., Y.Z. and L.Z.; Supervision, Y.Z.; Project administration, Y.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

Author Hao Wu was employed by the company China Southern Power Grid Energy Storage Co., Ltd. Author Junhuang Zhou was employed by the company Guangzhou Power Electrical Technology Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Pérez-Díaz, J.I.; Chazarra, M.; García-González, J.; Cavazzini, G.; Stoppato, A. Trends and challenges in the operation of pumped-storage hydropower plants. Renew. Sustain. Energy Rev. 2015, 44, 767–784. [Google Scholar] [CrossRef]
  2. Qiu, Y.; Li, Q.; Ai, Y.; Chen, W.; Benbouzid, M.; Liu, S.; Gao, F. Two-stage distributionally robust optimization-based coordinated scheduling of integrated energy system with electricity-hydrogen hybrid energy storage. Prot. Control Mod. Power Syst. 2023, 8, 33. [Google Scholar] [CrossRef]
  3. Alturki, F.A.; Awwad, E.M. Sizing and cost minimization of standalone hybrid wt/pv/biomass/pump-hydro storage-based energy systems. Energies 2021, 14, 489. [Google Scholar] [CrossRef]
  4. Maio, M.; Marrasso, E.; Roselli, C.; Sasso, M.; Fontana, N.; Marini, G. An innovative approach for optimal selection of pumped hydro energy storage systems to foster sustainable energy integration. Renew. Energy 2024, 227, 120533. [Google Scholar] [CrossRef]
  5. Vilanova, M.R.N.; Flores, A.T.; Balestieri, J.A.P. Pumped hydro storage plants: A review. J. Braz. Soc. Mech. Sci. Eng. 2020, 42, 415. [Google Scholar] [CrossRef]
  6. Hunt, J.D.; Zakeri, B.; Lopes, R.; Barbosa, P.S.F.; Nascimento, A.; de Castro, N.J.; Brandão, R.; Schneider, P.S.; Wada, Y. Existing and new arrangements of pumped-hydro storage plants. Renew. Sustain. Energy Rev. 2020, 129, 109914. [Google Scholar] [CrossRef]
  7. Blakers, A.; Stocks, M.; Lu, B.; Cheng, C. A review of pumped hydro energy storage. Prog. Energy 2021, 3, 022003. [Google Scholar] [CrossRef]
  8. Kong, L.; Nian, H. Fault detection and location method for mesh-type DC microgrid using pearson correlation coefficient. IEEE Trans. Power Deliv. 2020, 36, 1428–1439. [Google Scholar] [CrossRef]
  9. Jia, K.; Yang, Z.; Zheng, L.; Zhu, Z.; Bi, T. Spearman correlation-based pilot protection for transmission line connected to PMSGs and DFIGs. IEEE Trans. Ind. Inform. 2020, 17, 4532–4544. [Google Scholar] [CrossRef]
  10. Wang, Q.; Zhang, X.; Yi, C.; Li, Z.; Xu, D. A novel shared energy storage planning method considering the correlation of renewable uncertainties on the supply side. IEEE Trans. Sustain. Energy 2022, 13, 2051–2063. [Google Scholar] [CrossRef]
  11. Ma, Y.; Zhao, W.; Meng, M.; Zhang, Q.; She, Q.; Zhang, J. Cross-subject emotion recognition based on domain similarity of EEG signal transfer learning. IEEE Trans. Neural Syst. Rehabil. Eng. 2023, 31, 936–943. [Google Scholar] [CrossRef]
  12. Zhang, L.; Lu, D.; Wang, X. Measuring and testing interdependence among random vectors based on Spearman’s ρ and Kendall’s τ. Comput. Stat. 2020, 35, 1685–1713. [Google Scholar] [CrossRef]
  13. Bui, D.T.; Tsangaratos, P.; Nguyen, V.T.; Van Liem, N.; Trinh, P.T. Comparing the prediction performance of a Deep Learning Neural Network model with conventional machine learning models in landslide susceptibility assessment. Catena 2020, 188, 104426. [Google Scholar] [CrossRef]
  14. Zhang, J.; Zhang, Y.; Song, J.; Cheng, L.; Paul, P.K.; Gan, R.; Shi, X.; Luo, Z.; Zhao, P. Large-scale baseflow index prediction using hydrological modelling, linear and multilevel regression approaches. J. Hydrol. 2020, 585, 124780. [Google Scholar] [CrossRef]
  15. Chen, Y.; Gan, M.; Pan, S.; Pan, H.; Zhu, X.; Tao, Z. Application of auto-regressive (AR) analysis to improve short-term prediction of water levels in the Yangtze estuary. J. Hydrol. 2020, 590, 125386. [Google Scholar] [CrossRef]
  16. Kaur, J.; Parmar, K.S.; Singh, S. Autoregressive models in environmental forecasting time series: A theoretical and application review. Environ. Sci. Pollut. Res. 2023, 30, 19617–19641. [Google Scholar] [CrossRef]
  17. Mannering, F.; Bhat, C.R.; Shankar, V.; Abdel-Aty, M. Big data, traditional data and the tradeoffs between prediction and causality in highway-safety analysis. Anal. Methods Accid. Res. 2020, 25, 100113. [Google Scholar] [CrossRef]
  18. Sun, Y.; Ding, S.; Zhang, Z.; Jia, W. An improved grid search algorithm to optimize SVR for prediction. Soft Comput. 2021, 25, 5633–5644. [Google Scholar] [CrossRef]
  19. Gholizadeh, M.; Jamei, M.; Ahmadianfar, I.; Pourrajab, R. Prediction of nanofluids viscosity using random forest (RF) approach. Chemom. Intell. Lab. Syst. 2020, 201, 104010. [Google Scholar] [CrossRef]
  20. Liang, W.; Luo, S.; Zhao, G.; Wu, H. Predicting hard rock pillar stability using GBDT, XGBoost, and LightGBM algorithms. Mathematics 2020, 8, 765. [Google Scholar] [CrossRef]
  21. Chen, L.; Wu, T.; Wang, Z.; Lin, X.; Cai, Y. A novel hybrid BPNN model based on adaptive evolutionary Artificial Bee Colony Algorithm for water quality index prediction. Ecol. Indic. 2023, 146, 109882. [Google Scholar] [CrossRef]
  22. Williams, C.G.; Ojuri, O.O. Predictive modelling of soils’ hydraulic conductivity using artificial neural network and multiple linear regression. SN Appl. Sci. 2021, 3, 152. [Google Scholar] [CrossRef]
  23. Zhang, S.; Yu, J.J. Bayesian Deep Learning for Dynamic Power System State Prediction Considering Renewable Energy Uncertainty. J. Mod. Power Syst. Clean Energy 2022, 10, 913–922. [Google Scholar] [CrossRef]
  24. Liang, Y.; Ren, Y.; Yu, J.; Zha, W. Current Trajectory Image-Based Protection Algorithm for Transmission Lines Connected to MMC-HVDC Stations Using CA-CNN. Prot. Control Mod. Power Syst. 2023, 8, 6. [Google Scholar] [CrossRef]
  25. Sun, K.; Liu, D.; Cui, L. Rotating machinery fault diagnosis based on optimized Hilbert curve images and a novel bi-channel CNN with attention mechanism. Meas. Sci. Technol. 2023, 34, 125022. [Google Scholar] [CrossRef]
  26. Fan, H.; Xue, C.; Ma, J.; Cao, X.; Zhang, X. A novel intelligent diagnosis method of rolling bearing and rotor composite faults based on vibration signal-to-image mapping and CNN-SVM. Meas. Sci. Technol. 2023, 34, 044008. [Google Scholar]
  27. Li, S.; Cao, D.; Hu, W.; Huang, Q.; Chen, Z.; Blaabjerg, F. Multi-energy Management of Interconnected Multi-microgrid System Using Multi-agent Deep Reinforcement Learning. J. Mod. Power Syst. Clean Energy 2023, 11, 1606–1617. [Google Scholar] [CrossRef]
  28. Chen, H.; Meng, W.; Li, Y.; Xiong, Q. An anti-noise fault diagnosis approach for rolling bearings based on multiscale CNN-LSTM and a deep residual learning model. Meas. Sci. Technol. 2023, 34, 045013. [Google Scholar] [CrossRef]
  29. Zhang, H.; Sun, H.; Kang, L.; Zhang, Y.; Wang, L.; Wang, K. Prediction of Health Level of Multiform Lithium Sulfur Batteries Based on Incremental Capacity Analysis and an Improved LSTM. Prot. Control Mod. Power Syst. 2024, 9, 21–31. [Google Scholar] [CrossRef]
  30. Fu, G.; Wei, Q.; Yang, Y.; Li, C. Bearing fault diagnosis based on CNN-BiLSTM and residual module. Meas. Sci. Technol. 2023, 34, 125050. [Google Scholar] [CrossRef]
  31. Jin, R.; Chen, Z.; Wu, K.; Wu, M.; Li, X.; Yan, R. Bi-LSTM-based two-stream network for machine remaining useful life prediction. IEEE Trans. Instrum. Meas. 2022, 71, 3511110. [Google Scholar] [CrossRef]
  32. Kim, J.; Oh, S.; Kim, H.; Choi, W. Tutorial on time series prediction using 1D-CNN and BiLSTM: A case example of peak electricity demand and system marginal price prediction. Eng. Appl. Artif. Intell. 2023, 126, 106817. [Google Scholar] [CrossRef]
  33. Zhang, C.; Chen, M.; Zhang, Y.; Deng, W.; Gong, Y.; Zhang, D. Partial discharge pattern recognition algorithm of overhead covered conductors based on feature optimization and bidirectional LSTM-GRU. IET Gener. Transm. Distrib. 2024, 18, 680–693. [Google Scholar] [CrossRef]
  34. Dairi, A.; Harrou, F.; Khadraoui, S.; Sun, Y. Integrated multiple directed attention-based deep learning for improved air pollution forecasting. IEEE Trans. Instrum. Meas. 2021, 70, 3520815. [Google Scholar] [CrossRef]
  35. Brauwers, G.; Frasincar, F. A general survey on attention mechanisms in deep learning. IEEE Trans. Knowl. Data Eng. 2021, 35, 3279–3298. [Google Scholar] [CrossRef]
  36. Ziyabari, S.; Zhao, Z.; Du, L.; Biswas, S.K. Multi-branch resnet-transformer for short-term spatio-temporal solar irradiance forecasting. IEEE Trans. Ind. Appl. 2023, 59, 5293–5303. [Google Scholar] [CrossRef]
Figure 1. An operating schematic of a pumped storage hydropower plant.
Figure 1. An operating schematic of a pumped storage hydropower plant.
Energies 18 03057 g001
Figure 2. Data samples for upper (left) and lower (right) reservoir capacities.
Figure 2. Data samples for upper (left) and lower (right) reservoir capacities.
Energies 18 03057 g002
Figure 3. The structural framework of the Multi-Branch Attention–CNN–BiLSTM forecast model.
Figure 3. The structural framework of the Multi-Branch Attention–CNN–BiLSTM forecast model.
Energies 18 03057 g003
Figure 4. The structural framework of LSTM (upper) and BiLSTM (lower).
Figure 4. The structural framework of LSTM (upper) and BiLSTM (lower).
Energies 18 03057 g004
Figure 5. The structural framework of the SE attention mechanism in the CNN.
Figure 5. The structural framework of the SE attention mechanism in the CNN.
Energies 18 03057 g005
Figure 6. The branch network structure.
Figure 6. The branch network structure.
Energies 18 03057 g006
Figure 7. The distribution diagram of upper reservoir capacity forecast results.
Figure 7. The distribution diagram of upper reservoir capacity forecast results.
Energies 18 03057 g007
Figure 8. The scatter diagram of upper reservoir capacity forecast results.
Figure 8. The scatter diagram of upper reservoir capacity forecast results.
Energies 18 03057 g008
Figure 9. The distribution diagram of lower reservoir capacity forecast results.
Figure 9. The distribution diagram of lower reservoir capacity forecast results.
Energies 18 03057 g009
Figure 10. The scatter diagram of lower reservoir capacity forecast results.
Figure 10. The scatter diagram of lower reservoir capacity forecast results.
Energies 18 03057 g010
Figure 11. The 3 σ ReLu historgram of residuals for upper and lower reservoir capacities.
Figure 11. The 3 σ ReLu historgram of residuals for upper and lower reservoir capacities.
Energies 18 03057 g011
Table 1. Data types and units.
Table 1. Data types and units.
Data TypeUnitData TypeUnit
U01 powerkWU01 speedHz
U02 powerkWU02 speedHz
U03 powerkWU03 speedHz
U04 powerkWU04 speedHz
Upper reservoir level 1mUpper reservoir level 2m
Lower reservoir level 1mLower reservoir level 2m
Upper reservoir capacity m 3 Lower reservoir capacity m 3
Table 2. The Spearman correlation coefficients of each variable with the upper and lower reservoirs.
Table 2. The Spearman correlation coefficients of each variable with the upper and lower reservoirs.
Data TypeSpearman Correlation Coefficient
Upper CapacityLower CapacityAverage Value
U 01 power0.43520.21250.3239
U 01 speed0.02570.01860.0222
U 02 power0.04020.01450.0273
U 02 speed0.09040.05940.0749
U 03 power0.45030.16400.3071
U 03 speed0.17430.06020.1173
U 04 power0.18930.13830.1638
U 04 speed0.21310.21640.2148
Upper level 10.71810.26050.4893
Upper level 20.61040.27260.4415
Lower level 10.45370.49330.4735
Lower level 20.51640.38040.4484
Upper capacity1.00000.22130.6107
Lower capacity0.22131.00000.6107
Table 3. Multi-Branch Attention–CNN–BiLSTM model parameters.
Table 3. Multi-Branch Attention–CNN–BiLSTM model parameters.
ParameterValue
Number of CNN1 channels64
Number of CNN2 channels32
Number of CNN3 channels32
Number of FC1 channels for SE4
Number of FC2 channels for SE32
Neuron count of BiLSTM hidden layer32
Weighted argument[0.2444, 0.3969, 0.3587]
Table 4. Forecast of upper reservoir capacity in different models.
Table 4. Forecast of upper reservoir capacity in different models.
Model R 2 MAPE /% RMSE / m 3
BPNN0.74925.117148.5605
LSTM0.78764.844244.6952
BiLSTM0.79534.815343.8756
CNN-BiLSTM0.85333.883733.9849
SE-CNN-BiLSTM0.91142.862828.8582
Multi-Branch Attention–CNN–BiLSTM (ours)0.92312.885326.8869
Table 5. Forecast of lower reservoir capacity in different models.
Table 5. Forecast of lower reservoir capacity in different models.
Model R 2 MAPE /% RMSE / m 3
BPNN0.60779.097368.0587
LSTM0.64807.741464.4664
BiLSTM0.69667.673559.8526
CNN-BiLSTM0.72786.828351.8828
SE-CNN-BiLSTM0.79495.888749.2135
Multi-Branch Attention–CNN–BiLSTM (ours)0.82425.425145.5623
Table 6. Average forecast of upper and lower reservoir capacities in different models.
Table 6. Average forecast of upper and lower reservoir capacities in different models.
Model R 2 MAPE /% RMSE / m 3 Time/s
BPNN0.67857.107258.30960.0190
LSTM0.71786.292854.58080.0084
BiLSTM0.74606.244451.86410.0098
CNN-BiLSTM0.79065.356042.93390.0211
SE-CNN-BiLSTM0.85314.375839.03590.0217
Multi-Branch Attention–CNN–BiLSTM (ours)0.87374.155236.22460.0525
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gong, Y.; Wu, H.; Zhou, J.; Zhang, Y.; Zhang, L. Hybrid Multi-Branch Attention–CNN–BiLSTM Forecast Model for Reservoir Capacities of Pumped Storage Hydropower Plant. Energies 2025, 18, 3057. https://doi.org/10.3390/en18123057

AMA Style

Gong Y, Wu H, Zhou J, Zhang Y, Zhang L. Hybrid Multi-Branch Attention–CNN–BiLSTM Forecast Model for Reservoir Capacities of Pumped Storage Hydropower Plant. Energies. 2025; 18(12):3057. https://doi.org/10.3390/en18123057

Chicago/Turabian Style

Gong, Yu, Hao Wu, Junhuang Zhou, Yongjun Zhang, and Langwen Zhang. 2025. "Hybrid Multi-Branch Attention–CNN–BiLSTM Forecast Model for Reservoir Capacities of Pumped Storage Hydropower Plant" Energies 18, no. 12: 3057. https://doi.org/10.3390/en18123057

APA Style

Gong, Y., Wu, H., Zhou, J., Zhang, Y., & Zhang, L. (2025). Hybrid Multi-Branch Attention–CNN–BiLSTM Forecast Model for Reservoir Capacities of Pumped Storage Hydropower Plant. Energies, 18(12), 3057. https://doi.org/10.3390/en18123057

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop