Next Article in Journal
Artificial Intelligence and Sustainable Practices in Coastal Marinas: A Comparative Study of Monaco and Ibiza
Previous Article in Journal
LEED v4 Adoption Patterns and Regional Variations Across US-Based Projects
Previous Article in Special Issue
Prediction and Control of Existing High-Speed Railway Tunnel Deformation Induced by Shield Undercrossing Based on BO-XGboost
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Bidirectional Gated Recurrent Unit (BiGRU)-Based Model for Concrete Gravity Dam Displacement Prediction

1
Guangxi Youjiang Water Resources Development Co., Ltd., Nanning 530022, China
2
School of Water Resources and Hydropower Engineering, Wuhan University, Wuhan 430072, China
*
Author to whom correspondence should be addressed.
Sustainability 2025, 17(16), 7401; https://doi.org/10.3390/su17167401
Submission received: 6 June 2025 / Revised: 6 July 2025 / Accepted: 11 July 2025 / Published: 15 August 2025

Abstract

Dam displacement serves as a critical visual indicator for assessing structural integrity and stability in dam engineering. Data-driven displacement forecasting has become essential for modern dam safety monitoring systems, though conventional approaches—including statistical models and basic machine learning techniques—often fail to capture comprehensive feature representations from multivariate environmental influences. To address these challenges, a bidirectional gated recurrent unit (BiGRU)-enhanced neural network is developed, incorporating sliding window mechanisms to model time-dependent hysteresis characteristics. The BiGRU’s architecture systematically integrates historical temporal patterns through overlapping window segmentation, enabling dual-directional sequence processing via forward–backward gate structures. Validated with four instrumented measurement points from a major concrete gravity dam, the proposed model exhibits significantly better performance against three widely used recurrent neural network benchmarks in displacement prediction tasks. These results confirm the model’s capability to deliver high-fidelity displacement forecasts with operational stability, establishing a robust framework for infrastructure health monitoring applications.

1. Introduction

Dams serve as critical infrastructure for water resource management, hydropower generation, and ecological balance. Consequently, ensuring their long-term operational safety has become a global priority and a cornerstone of sustainable development for resilient infrastructure. Dam displacement is a primary indicator of structural health, but its prediction is complicated by the interplay of hydrostatic pressure, thermal effects, and irreversible material aging. Consequently, developing accurate displacement prediction models that fuse physical mechanisms with data-driven insights remains a key research frontier in dam engineering, directly supporting the sustainable stewardship of vital water resources and the communities they serve.
Displacement prediction methodologies have evolved from physically interpretable statistical models to sophisticated machine learning frameworks. The cornerstone of the statistical approach is the Hydrostatic–Temperature–Time (HTT) model, which leverages regression analysis to quantify deformation components [1]. This framework has been progressively enhanced to handle complex geometries in high dams [2], to incorporate cold-climate effects [3], and to couple spatial and seepage dynamics [4]. The recent paradigm shift towards machine learning [5] has unlocked new capabilities. For instance, Su et al. [6] pioneered hybrid frameworks combining support vector machines (SVM) with finite element simulations, which Wang et al. [7] later complemented by incorporating Gaussian process regression. Tree-based ensembles have also been prominent; Alazar et al. [8] leveraged gradient-boosted regression trees for displacement and seepage modeling, Yang et al. [9] proposed a hybrid XGBoost–ANN model for residual prediction, and Su et al. [10] advanced random forest (RF) models with sliding time windows.
Within the domain of neural networks, significant progress has been made. Huang et al. [11] proposed a dual-attention long short-term memory (DALSTM) model to effectively capture thermal effects. To enhance generalizability, Xu et al. [12] synergized LSTM with wavelet decomposition, while to improve predictive robustness, Li et al. [13] developed a one-dimensional residual network and LSTM (DRLSTM) architecture. Further innovations in neural architecture have also been explored. Peng et al. [14] integrated graph convolutional networks (GCN) with attention mechanisms for distributed sensor feature extraction, and Kang et al. [15] applied extreme learning machines (ELM) with modified activation functions to gravity dam health monitoring.
Among these methods, the Gated Recurrent Unit (GRU) is particularly adept at modeling long-term dependencies while remaining computationally efficient. Its efficacy in dam engineering has been validated in various specialized models. Yuan et al. [16] proposed a VMD–TSVR–GRU model to account for non-stationarity in displacement data. Lu et al. [17] validated an Inception–ResNet–GRU model on the Ertan Dam, achieving significant RMSE reduction. Concurrently, Xu et al. [18] utilized a CNN–GRU model with spatial pooling to predict zonal displacement clusters. Despite these advancements, challenges in mitigating overfitting and ensuring comprehensive feature extraction from limited data remain [19]. Furthermore, most recurrent models are unidirectional, struggling to fully capture the complex hysteresis phenomena where structural responses lag behind load variations [20,21].
To address the limitations of unidirectional models, especially in capturing the pronounced posteriority (time-lag effect) of temperature on dam deformation, this paper proposes an enhanced predictive model based on a Bidirectional GRU (BiGRU) network. This posteriority signifies that a dam’s structural response is a function of the temperature history over an extended preceding period, rather than just the instantaneous temperature. Our framework explicitly tackles this by incorporating a sliding window mechanism for feature engineering [13,22]. This technique provides the model not with a single snapshot in time, but with a sequence of recent historical data, thereby embedding the necessary temporal context to model the lag. The BiGRU architecture then processes these entire sequences in both forward and backward directions, allowing it to effectively learn the complex dependencies and time-lag patterns inherent in the data. The primary contribution of this work is the development and validation of this integrated framework on a 130 m concrete gravity dam. By providing more reliable foresight into structural behavior, this work offers a critical tool for proactive maintenance and risk management, thereby enhancing the long-term sustainability and resilience of critical water infrastructure.
The remainder of this paper is structured as follows: Section 2 details the methodology. Section 3 introduces the case study and evaluation metrics, and details the optimization of key parameters, including the sliding window size and the model’s hyperparameters. Section 4 presents the experimental results and model comparisons. Finally, Section 5 discusses the findings and outlines future research directions.

2. Methodology

As illustrated in Figure 1, this framework is conceptually guided by the classical hydrostatic–seasonal–time (HTT) model’s principles and implemented using a powerful bidirectional gated recurrent unit (BiGRU) neural network. The implementation comprises three sequential phases: (1) Feature engineering guided by the HTT framework: Instead of performing a statistical decomposition, this phase focuses on selecting and engineering a comprehensive feature set. We utilize the raw monitoring variables that correspond to the HTT components (e.g., water levels for the hydrostatic effect, temperatures for the thermal effect). These are then enhanced with additional time-series features (e.g., lagged values and rolling statistics of the displacement) to holistically capture the time-dependent effects, such as material creep and seasonal patterns. (2) Sequence construction via sliding window: The resulting multi-feature time series is then partitioned using a sliding window technique. This creates discrete yet overlapping intervals, thereby capturing sequential dependencies and inherent time-lag characteristics intrinsic to dam deformation mechanisms; (3) Predictive modeling with a BiGRU: Finally, the preprocessed sequences are fed into the BiGRU architecture. The network leverages forward–backward propagation to model non-linear interactions between historical and future states, ultimately generating long-term deformation forecasts alongside corresponding statistical evaluation metrics (RMSE, MAE, R2). To ensure operational validity, the framework is rigorously validated through an engineering case study of a 130 m-high concrete gravity dam with over 18 years of operational history. Four strategically positioned monitoring points are utilized to evaluate spatiotemporal prediction accuracy across distinct dam zones.

2.1. Hydraulic–Thermal Transfer (HTT) Statistical Model

The Hydraulic Structure Temperature (HST) model utilizes harmonic functions to represent thermal variations yet fails to adequately explain thermally induced displacements arising from both short-term ambient temperature fluctuations and multi-year climatic variations. Conversely, the HTT model directly derives its thermal component from in situ temperature measurements, achieving a superior accuracy in modeling real-world dam monitoring data. The HTT framework decomposes concrete dam displacements (δ) into three constitutive terms: the static hydrostatic pressure component (δH), the thermal effects component (δT), and the time-dependent creep component (δθ), expressed as:
δ = δ H + δ T + δ θ  
The hydrostatic component (δH) is governed by the reservoir elevation, expressed as:
δ H = i = 1 n a i H i
where ai are the regression coefficients; H denotes the reservoir elevation, with the exponent n = 3 for gravity dams and n = 4 for arch dams.
The thermal component δT characterizes thermally induced displacements from concrete and foundation temperature fluctuations, calculated using measured temperatures:
δ T = i = 1 m b i T i
where bi are the thermal regression coefficients, m indicates the number of temperature sensors, and Ti represents the measured temperature values.
The time-dependent component represents concrete creep deformation:
δ θ = c 1 θ + c 2 ln θ
θ = t / 100
where c1 and c2 are the time-effect coefficients, and t denotes the cumulative operational days post-construction.
The aforementioned equations represent the classical statistical formulation of the HTT model, which provides a robust theoretical basis for understanding the primary factors of dam displacement. However, this traditional approach often assumes linear relationships and may struggle to capture the full spectrum of complex, non-linear dynamics inherent in dam behavior.
Therefore, in our proposed framework, we pivot from this statistical decomposition. Instead of using the calculated components (δH, δT, δθ) as inputs, we leverage the fundamental principles of the HTT model to guide our feature engineering. The raw physical variables that constitute these components (i.e., water levels, temperatures, and time-related markers) are fed directly into a deep learning network. This data-driven methodology allows the BiGRU model to learn the underlying non-linear relationships directly, offering the potential for a more accurate and comprehensive prediction.

2.2. BiGRU Model

Traditional Recurrent Neural Networks (RNNs) are constrained by vanishing gradient and exploding gradient phenomena, which impair their performance on long-sequence modeling. The GRU addresses these limitations through gating mechanisms—a reset gate (rt) for short-term pattern extraction and an update gate (zt) for long-term state retention [23]. Figure 2 illustrates its computational graph at a single time step:
xt denotes the input vector of the time point, zt denotes the update gate, rt denotes the reset gate, ht−1 denotes the hidden state, h ˜ denotes the candidate hidden state, and ht denotes the updated hidden state. The sigmoid function (σ) generates gate activations (0–1), controlling information flow:
r t = σ ( W r [ h t 1 , x t ] )
z t = σ ( W z [ h t 1 , x t ] )
h ˜ t = tanh ( W h ˜ [ r t h t 1 , x t ] )
h t = ( 1 z t ) h t 1 + z t h ˜ t
where denotes element-wise multiplication. The update gate interpolates between prior state retention (zt → 0) and new state adoption (zt → 1), while the reset gate rt modulates historical information integration.
The BiGRU extends the GRU’s representational power through dual-directional sequence processing. By jointly analyzing forward (past-to-future) and reverse (future-to-past) contexts, the BiGRU captures comprehensive temporal patterns, demonstrating superior performance in sequential modeling tasks. As shown in Figure 3, the BiGRU framework integrates two GRU layers. The architecture implements bidirectional temporal processing through two complementary units: the forward GRU extracts features by propagating hidden states chronologically from timestep t = 1 to t = T, capturing historical dependencies, while the backward GRU operates inversely, propagating from t = T to t = 1 to model future-influenced patterns. Hidden state outputs from both directions are concatenated, enabling a synergistic integration of past and future contextual information for enhanced sequential representation learning.
At each time step t, the forward hidden state h t f w and backward hidden state h t b w are computed independently as follows:
h t f w = GRU ( x t , h t 1 f w )
h t b w = GRU ( x t , h t + 1 b w )
The final output yt concatenates both states as:
y t = [ h t f w ; h t b w ]
This bidirectional architecture enables the synergistic learning of historical influences and future trends, providing a holistic representation of time-series dynamics.

2.3. Rolling Window Feature Engineering

To address delayed hydrological effects in dam monitoring, we employ rolling window aggregation to inject temporally shifted context into the HTT framework. Wτ = [tτ, t − 1] defines a backward-looking window of days. The method computes six statistical descriptors over Wτ as:
s t = μ , x ˜ , σ , m a x , m i n , I Q R W τ
where μ denotes the mean, x ˜ denotes the median, σ denotes the standard deviation, max denotes the max, min denotes the min, and IQR denotes the interquartile range. As diagrammed in Figure 4, for τ = 3, these features derive from the interval [t − 3, t − 1], creating a 72 h contextual prior. The engineered features st are concatenated with raw inputs xt as:
x ^ t = x t s t
where denotes the feature concatenation. The enriched input x ^ t feeds into the BiGRU layer, enabling a simultaneous learning of instantaneous signals and historical trends. Explicit feature engineering enhances interpretability by quantifying how statistical properties of past observations influence current predictions.

3. Case Study

3.1. Project Description

This case study focuses on the main dam of the Baise Water Control Project in Guangxi, China. The front view of the dam is illustrated in Figure 5. The Baise Water Control Project began construction in 2001 and was completed in 2006. Table 1 presents the monitored hydraulic design parameters of the Baise Water Control Project.
The dam complex consists of the main dam, a hydropower station, two auxiliary dams, and the first phase of the navigation building, arranged in a decentralized layout. The main dam is a full-section roller-compacted concrete dam with a maximum height of 130 m, a crest length of 720 m, and 27 monoliths. The spillway section is 88 m long, and equipped with four overflow surface gates and three flood discharge mid-level outlets, with a maximum discharge capacity of 13,737 m3/s.
Based on the actual monitoring data and spatial distribution of the monitoring points, four representative monitoring points were selected from the distinct zones of the dam to serve as training samples. These points (TC3-1, TC3-11, TC3-12, and TC3-17) are vertical displacement sensors located at the dam crest. Their specific locations are shown in Figure 6 (schematic diagram of dam cross-section layout).

3.2. Data Collection and Preprocessing

The study analyzed daily monitoring data collected between April 2022 and August 2024. Figure 7 presents the temporal variations in the environmental parameters, including reservoir level and temperature. The reservoir level fluctuated seasonally, peaking at 223 m during flood seasons and reaching a minimum operational level of 202 m in dry periods. To mitigate flood risks associated with the summer monsoon rainfall, annual drawdown operations reduced the reservoir to its flood control level (202 m) from February to April. Post-monsoon replenishment restored the reservoir to its normal operating level (215 m) to support winter hydropower generation and irrigation demands.
Figure 8 details the vertical displacement patterns recorded at four crest monitoring points on the main dam. Globally, displacement data across the monitoring points display a seasonal trend: a downward trajectory from summer (June–August) to winter (December–February), followed by a gradual ascent to peak values in spring (March–May). Comparative analysis revealed a substantial heterogeneity in the initial displacements among the four monitoring points, with distinct variation ranges and amplitudes. Notably, discontinuous displacement patterns were identified in all monitoring datasets, attributable to the combined effects of natural factors (e.g., temperature fluctuations) and anthropogenic activities. These irregularities complicate the model’s capacity to discern dominant features within the high-dimensional data space.
To ensure optimal convergence and model robustness, a bifurcated data conditioning regimen was strategically implemented. The input features underwent Z-score standardization (Equation (15)), a process that normalizes each feature to a zero mean and unit variance. This standardization is critical for mitigating the influence of disparate feature scales and enhancing robustness to outliers, which are common in real-world environmental data. Concurrently, the target variable was subjected to Min–Max normalization (Equation (16)), a transformation that confines its values to the bounded [0, 1] interval. This latter procedure stabilizes the predictive target for the network’s output layer, thereby facilitating a more stable and efficient gradient descent during training. This dual-pronged strategy is predicated on the rationale that conditioning inputs and outputs differentially according to their roles maximizes the efficacy of the learning algorithm.
x standard = x μ σ
where x represents the original value of the feature, while μ and σ denote the mean and standard deviation of that feature, respectively.
x norm = x min ( x ) max ( x ) min ( x )
where x refers to the original data; xnorm refers to the normalized data.

3.3. Evaluation Metrics

Three evaluation metrics were employed to assess the predictive performance of the model: Root Mean Square Error (RMSE), Mean Absolute Error (MAE), and Coefficient of Determination (R2). The corresponding mathematical expressions are defined as:
R M S E = 1 n i = 1 n y i y ^ i 2
M A E = 1 n i = 1 n | y i y ^ i |
R 2 = 1 i = 1 n y i y ^ i 2 i = 1 n y i y ¯ 2
where y ^ i denotes the predicted values, yi represents the actual values, y ¯ is the mean of all samples, and n is the total number of samples.

3.4. Model Implementation and Hyperparameter Tuning

After defining the data, evaluation metrics, and the general BiGRU architecture, the next crucial phase is to implement the model and systematically optimize its parameters for the specific context of this case study. Generic model architecture rarely yields optimal results; therefore, a two-stage tuning process was employed to ensure the model’s configuration was best suited for the predictive task. First, we optimized the key feature engineering parameter—the rolling window size. Second, with the optimal features established, we proceeded to tune the internal hyperparameters of the BiGRU model itself. This data-driven approach ensures the model’s robustness and maximizes its predictive performance.

3.4.1. Optimization of Rolling Window Size

The construction of the time-series features using a rolling window is fundamental to providing the model with temporal context. The size of this window dictates a critical trade-off: a smaller window is more sensitive to recent fluctuations but can be susceptible to noise, while a larger window provides smoother features at the cost of increased lag.
To empirically determine the optimal window size, a comparative experiment was conducted. We evaluated a range of window sizes, including 2, 3, 4, 5, 7, and 10. For each size, the full data processing and model training pipeline was executed while holding all other model hyperparameters (e.g., neuron counts, learning rate) constant at a baseline value. The model’s performance was assessed using the RMSE, MAE, and R2 metrics, as defined in Section 3.3.
The comprehensive results of this experiment are presented in Table 2. The analysis indicates that a window size of three consistently yielded a superior or highly competitive performance across the majority of the datasets. It achieved a robust balance between capturing meaningful short-term dynamics and maintaining feature stability. Therefore, a rolling window size of three was selected for all feature engineering in the subsequent stages of this study.

3.4.2. BiGRU Model Hyperparameter Optimization

With the input features defined by the optimal rolling window size, the next step was to optimize the BiGRU model’s internal architecture. This process is essential for ensuring the model’s robustness and tailoring its complexity to the specific dataset, thereby preventing over- or under-fitting.
For this task, we employed the Keras Tuner library, a dedicated tool for hyperparameter optimization. We utilized the Hyperband algorithm, an efficient search strategy, to find the best combination of hyperparameters by minimizing the validation loss (val_loss). The search space was defined to include the model’s most critical parameters: Units in the first BiGRU layer—an integer value between 32 and 256 (step size: 32); units in the second BiGRU layer—an integer value between 16 and 128 (step size: 16); dropout rates—a floating-point value between 0.2 and 0.5 (step size: 0.1); and learning rate—a selection from the set {0.01, 0.001, 0.0001}.
The optimization process automatically trained and evaluated numerous model configurations. The best-performing set of hyperparameters identified through this search is detailed in Table 3. All final model training, evaluation, and analysis presented in Section 4 were conducted using this optimized hyperparameter configuration.

4. Results

4.1. Experimental Setup

To ensure the transparency and reproducibility of our research, it is essential to document the precise configuration of the experimental environment. Accordingly, all pertinent details regarding the hardware platform, software dependencies, and the final model hyperparameters employed in this study are systematically summarized in Table 4. This information includes the specifications of the computing hardware, the versions of key libraries, and the optimal parameter values determined through the tuning processes described in the preceding sections.

4.2. Model Training and Convergence Analysis

To analyze the convergence and stability of the model during the training phase, the learning curves for the optimized BiGRU model at each of the four monitoring points are presented in Figure 9. A consistent and desirable training pattern is observed across all four subplots (a–d).
In each case, both the training loss and validation loss decrease sharply during the initial epoch, indicating that the model learns the primary data patterns rapidly. Subsequently, the curves flatten and converge to a stable, low value. Crucially, the validation loss curve closely tracks the training loss curve throughout the entire process, without any significant divergence. This proximity between the two curves provides strong evidence that the models achieved effective generalization without overfitting the training data. Furthermore, the varying number of total epochs in each subplot demonstrates the effective functioning of the Early Stopping mechanism, which terminated the training process optimally for each case once the validation loss ceased to improve.

4.3. Comparative Analysis of Model Performance

To benchmark the baseline model performance, four time-series forecasting architectures—LSTM, BiLSTM, GRU, and BiGRU—were trained on raw measured displacement data using a three-step sliding window with default hyperparameters. The model’s generalization capability was evaluated through long-term prediction accuracy and standard evaluation metrics. As the dam’s service duration extends, operational monitoring datasets exhibit a progressive temporal expansion.
The quantitative evaluation of the proposed method and baseline models for long-term dam displacement prediction is summarized in Table 5. The BiGRU algorithm consistently outperformed other architectures across all four monitoring points, demonstrating an enhanced generalization capacity and robust stability. This superiority stems from its dual-directional memory propagation mechanism, which adaptively weights historical context and future-inferred patterns—a critical advantage when processing sparse or asynchronous monitoring data. When using the correlation coefficients as the evaluation metric, the BiGRU exhibited statistically significant improvements over LSTM, BiLSTM, and GRU, with performance gains closely aligned with the complexity of each baseline’s temporal dependency modeling capability. These results demonstrate that the BiGRU is capable of extracting meaningful patterns from four distinct and weakly correlated datasets, validating its effectiveness in disentangling heterogeneous spatiotemporal couplings within dam systems.

4.4. Visual Analysis of Prediction Results

Figure 10 provides a comparative visualization of the long-term prediction performance between the proposed method and baseline models, validated against monitoring data collected from April to August 2024. The BiGRU model shows the strongest agreement with observational trends during the local temporal interval (April–July), achieving the lowest overall prediction error among all models. Its trajectory closely follows the phase-shifted deformation pulses induced by delayed reservoir seepage pressures—an effect that unidirectional models fail to capture effectively. All methods exhibit residual inaccuracies during the terminal prediction phase (August 2024), with errors systematically increasing compared with earlier stages. This divergence becomes more pronounced near hydrological transition periods, where conventional models struggle to reconcile rapid water level fluctuations with slow-moving creep deformations. The consistent performance degradation across models highlights the inherent challenges in maintaining predictive stability over extended horizons, underscoring the necessity of memory-gated architectures for effective error attenuation.
The observed difference in the magnitude of displacement fluctuations between the monitoring points can be attributed to the distinct structural characteristics and boundary conditions at their respective locations on the dam.
Points TC3-1 and TC3-17 are situated at the dam abutments, where the concrete structure is anchored into the bedrock of the valley sides. As shown in the profile diagram, these sections are thinner and have a smaller volume compared to the central monoliths. This lower thermal mass makes them more susceptible to ambient temperature variations, leading to more pronounced and frequent cycles of expansion and contraction. This heightened sensitivity to thermal effects is the primary reason for the larger displacement fluctuations observed at the dam’s flanks.
In contrast, points TC3-11 and TC3-12 are located on the crest of the massive central dam monoliths, directly above the powerhouse and spillway block. Due to their enormous volume, these sections possess significant thermal inertia, meaning their temperature changes much more slowly and is less influenced by short-term air temperature swings. Consequently, their displacement behavior is more stable and is predominantly governed by the more slowly varying hydrostatic pressure from the reservoir. Therefore, while their absolute displacement may be substantial, the fluctuations around their mean positions are less significant than those at the abutments.
The proposed BiGRU-based predictive framework demonstrates an exceptional generalization capability and interpretability, achieving a superior average performance across all four monitoring points with an R2 of 0.89, an MAE of 1.17 mm, and an RMSE of 1.70 mm. This performance represents a significant leap over conventional architectures; for instance, compared to the standard GRU model—the strongest of the benchmarks—our proposed model delivered a 20.2% reduction in RMSE and a 30.4% reduction in MAE. This quantitative superiority is visually substantiated by the analysis presented in Figure 11, where the BiGRU model’s residual box plots consistently exhibit a more compact and zero-centered error distribution. While the standard GRU showed competitive performance at a single point, the BiGRU model’s consistent high accuracy across all locations underscores its superior stability and robustness. These findings indicate that the synergistic integration of the bidirectional architecture and sliding window optimization enables the model to effectively extract latent spatiotemporal patterns, making it a promising solution for long-term displacement forecasting in concrete gravity dams.

5. Conclusions

In this study, we developed and validated an enhanced predictive framework for dam displacement by integrating a Bidirectional GRU (BiGRU) with optimized time-series feature engineering. The primary contributions and findings of this work are summarized as follows:
(1)
Novel hysteresis-aware framework: A predictive model was established that explicitly accounts for the hydraulic hysteresis effects in dam behavior. This was achieved by synergizing a BiGRU architecture, which captures the temporal dependencies from both past and future contexts, with a sliding window mechanism for dynamic feature extraction.
(2)
Systematic model optimization: The model’s robustness and configuration were not arbitrary but were determined through a rigorous, two-stage optimization process. A preliminary comparative experiment identified the optimal rolling window size (τ = 3) for feature engineering, followed by a systematic hyperparameter tuning process to ascertain the ideal network architecture, learning rate, and regularization parameters.
(3)
Superior quantitative performance: The optimized BiGRU model demonstrated superior predictive accuracy and generalization ability. Across all four monitoring points, it achieved an average R2 of 0.89 and an RMSE of 1.70 mm. This represents a substantial improvement over the strongest benchmark model (standard GRU), with a 20.2% reduction in RMSE and a 30.4% reduction in MAE, providing concrete evidence of the framework’s value.
(4)
Enhanced temporal feature representation: The success of the model is attributed to the BiGRU’s dual-path processing (forward and backward), which distills a more comprehensive feature representation from complex multivariate time-series data compared to unidirectional models. This establishes a new paradigm for enhancing dam safety assessments through bidirectional temporal learning.
Despite the model’s strong performance, certain limitations highlight avenues for future research. First, the framework’s sensitivity to displacement anomalies induced by operational fluctuations or sensor drifts warrants the development of integrated data-cleaning and anomaly detection protocols. Second, as a dam deforms as an integrated system, future work should prioritize multi-point data fusion strategies. Investigating cross-correlation analysis and employing advanced methodologies, like graph neural networks, could better quantify the spatiotemporal interdependencies among monitoring points, which is essential for modeling the holistic behavior of the entire structure. Finally, the validation in this study was limited to monitoring points on the dam crest due to limited data availability; future work should aim to verify the model’s performance across the entire dam profile by incorporating data from its middle and base sections. Successfully addressing these limitations will lead to more robust predictive systems, providing operators with critical tools for proactive maintenance and asset management, and thereby ensuring the long-term sustainability and resilience of vital water infrastructure.

Author Contributions

Conceptualization, X.H. and J.M.; methodology, J.M. and K.Y.; software, H.W.; validation, K.Y. and H.W.; formal analysis, K.Y. and X.H.; investigation, X.H.; resources, X.H.; data curation, X.H.; writing—original draft preparation, X.H. and J.M.; writing—review and editing, H.W. and K.Y.; visualization, H.W.; supervision, Y.L.; project administration, J.M.; funding acquisition, Y.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research has been partially supported by the Major Science and Technology Project of the Ministry of Water Resources of China (SKS-2022158).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author due to appropriate reasons.

Conflicts of Interest

Authors Xiaobing Huang and Jianxin Ma were employed by Guangxi Youjiang Water Resources Development Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Wang, S.; Gu, C.; Liu, Y.; Gu, H.; Xu, B.; Wu, B. Displacement Observation Data-Based Structural Health Monitoring of Concrete Dams: A State-of-Art Review. Structures 2024, 68, 107072. [Google Scholar] [CrossRef]
  2. Wang, S.; Xu, C.; Gu, C.; Su, H.; Wu, B. Hydraulic-Seasonal-Time-Based State Space Model for Displacement Monitoring of High Concrete Dams. Trans. Inst. Meas. Control 2021, 43, 3347–3359. [Google Scholar] [CrossRef]
  3. Gu, C.; Zhu, M.; Wu, Y.; Chen, B.; Zhou, F.; Chen, W. Multi-Output Displacement Health Monitoring Model for Concrete Gravity Dam in Severely Cold Region Based on Clustering of Measured Dam Temperature Field. Struct. Health Monit. 2023, 22, 3416–3436. [Google Scholar] [CrossRef]
  4. Liu, B.; Wei, B.; Li, H.; Mao, Y. Multipoint Hybrid Model for RCC Arch Dam Displacement Health Monitoring Considering Construction Interface and Its Seepage. Appl. Math. Model. 2022, 110, 674–697. [Google Scholar] [CrossRef]
  5. Su, H.; Wen, Z.; Sun, X.; Yang, M. Time-Varying Identification Model for Dam Behavior Considering Structural Reinforcement. Struct. Saf. 2015, 57, 1–7. [Google Scholar] [CrossRef]
  6. Su, H.; Li, X.; Yang, B.; Wen, Z. Wavelet Support Vector Machine-Based Prediction Model of Dam Deformation. Mech. Syst. Signal Process. 2018, 110, 412–427. [Google Scholar] [CrossRef]
  7. Wang, S.; Xu, C.; Liu, Y.; Wu, B. A Spatial Association-Coupled Double Objective Support Vector Machine Prediction Model for Diagnosing the Deformation Behaviour of High Arch Dams. Struct. Health Monit. 2022, 21, 945–964. [Google Scholar] [CrossRef]
  8. Salazar, F.; Toledo, M.Á.; Oñate, E.; Suárez, B. Interpretation of Dam Deformation and Leakage with Boosted Regression Trees. Eng. Struct. 2016, 119, 230–251. [Google Scholar] [CrossRef]
  9. Yang, X.; Xiang, Y.; Shen, G.; Sun, M. A Combination Model for Displacement Interval Prediction of Concrete Dams Based on Residual Estimation. Sustainability 2022, 14, 16025. [Google Scholar] [CrossRef]
  10. Su, Y.; Weng, K.; Lin, C.; Zheng, Z. An Improved Random Forest Model for the Prediction of Dam Displacement. IEEE Access 2021, 9, 9142–9153. [Google Scholar] [CrossRef]
  11. Huang, B.; Kang, F.; Li, J.; Wang, F. Displacement Prediction Model for High Arch Dams Using Long Short-Term Memory Based Encoder-Decoder with Dual-Stage Attention Considering Measured Dam Temperature. Eng. Struct. 2023, 280, 115686. [Google Scholar] [CrossRef]
  12. Xu, B.; Chen, Z.; Wang, X.; Bu, J.; Zhu, Z.; Zhang, H.; Wang, S.; Lu, J. Combined Prediction Model of Concrete Arch Dam Displacement Based on Cluster Analysis Considering Signal Residual Correction. Mech. Syst. Signal Process. 2023, 203, 110721. [Google Scholar] [CrossRef]
  13. Li, M.; Li, M.; Ren, Q.; Li, H.; Song, L. DRLSTM: A Dual-Stage Deep Learning Approach Driven by Raw Monitoring Data for Dam Displacement Prediction. Adv. Eng. Inform. 2022, 51, 101510. [Google Scholar] [CrossRef]
  14. He, P.; Pan, J.; Li, Y. Long-Term Dam Behavior Prediction with Deep Learning on Graphs. J. Comput. Des. Eng. 2022, 9, 1230–1245. [Google Scholar] [CrossRef]
  15. Kang, F.; Liu, J.; Li, J.; Li, S. Concrete Dam Deformation Prediction Model for Health Monitoring Based on Extreme Learning Machine. Struct. Control Health Monit. 2017, 24, e1997. [Google Scholar] [CrossRef]
  16. Yuan, D.; Gu, C.; Wei, B.; Qin, X.; Xu, W. A High-Performance Displacement Prediction Model of Concrete Dams Integrating Signal Processing and Multiple Machine Learning Techniques. Appl. Math. Model. 2022, 112, 436–451. [Google Scholar] [CrossRef]
  17. Lu, T.; Gu, C.; Yuan, D.; Zhang, K.; Shao, C. Deep Learning Model for Displacement Monitoring of Super High Arch Dams Based on Measured Temperature Data. Measurement 2023, 222, 113579. [Google Scholar] [CrossRef]
  18. Xu, B.; Chen, Z.; Su, H.; Zhang, H.; Zhu, Z. Interval Prediction Method for Dynamic Zoning of Displacements in Super-High Arch Dam under Significant Water Level Fluctuations. Struct. Health Monit. 2024. [Google Scholar] [CrossRef]
  19. Sun, S.; Wang, S.; Wei, Y. A New Ensemble Deep Learning Approach for Exchange Rates Forecasting and Trading. Adv. Eng. Inform. 2020, 46, 101160. [Google Scholar] [CrossRef]
  20. Li, M.; Si, W.; Ren, Q.; Song, L.; Liu, H. An Integrated Method for Evaluating and Predicting Long-Term Operation Safety of Concrete Dams Considering Lag Effect. Eng. Comput. 2021, 37, 2505–2519. [Google Scholar] [CrossRef]
  21. Ren, Q.; Li, M.; Song, L.; Liu, H. An Optimized Combination Prediction Model for Concrete Dam Deformation Considering Quantitative Evaluation and Hysteresis Correction. Adv. Eng. Inform. 2020, 46, 101154. [Google Scholar] [CrossRef]
  22. Xu, B.; Zhang, H.; Xia, H.; Song, D.; Zhu, Z.; Chen, Z.; Lu, J. A Multi-Level Prediction Model of Concrete Dam Displacement Considering Time Hysteresis and Residual Correction. Meas. Sci. Technol. 2025, 36, 15107. [Google Scholar] [CrossRef]
  23. Cho, K.; Van, M.B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning Phrase Representations Using RNN Encod-er-Decoder for Statistical Machine Translation. In Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 25–29 October 2014. [Google Scholar] [CrossRef]
Figure 1. Flowchart of the proposed model.
Figure 1. Flowchart of the proposed model.
Sustainability 17 07401 g001
Figure 2. Single-time-step computational graph of the GRU cell.
Figure 2. Single-time-step computational graph of the GRU cell.
Sustainability 17 07401 g002
Figure 3. Bidirectional GRU architecture.
Figure 3. Bidirectional GRU architecture.
Sustainability 17 07401 g003
Figure 4. Temporal coverage and statistical aggregation via rolling windows.
Figure 4. Temporal coverage and statistical aggregation via rolling windows.
Sustainability 17 07401 g004
Figure 5. Front view of Baise Water Dam.
Figure 5. Front view of Baise Water Dam.
Sustainability 17 07401 g005
Figure 6. Schematic diagram of dam cross-section layout.
Figure 6. Schematic diagram of dam cross-section layout.
Sustainability 17 07401 g006
Figure 7. Temporal evolution of environmental variables: (a) reservoir water level; (b) in situ temperature measurements.
Figure 7. Temporal evolution of environmental variables: (a) reservoir water level; (b) in situ temperature measurements.
Sustainability 17 07401 g007
Figure 8. Vertical displacement trends at four monitoring points (2022–2024).
Figure 8. Vertical displacement trends at four monitoring points (2022–2024).
Sustainability 17 07401 g008
Figure 9. Training and validation loss curves for the BiGRU model at the four monitoring points: (a) TC3-1, (b) TC3-11, (c) TC3-12, and (d) TC3-17.
Figure 9. Training and validation loss curves for the BiGRU model at the four monitoring points: (a) TC3-1, (b) TC3-11, (c) TC3-12, and (d) TC3-17.
Sustainability 17 07401 g009
Figure 10. Predictive performance comparison between BiGRU and three canonical time-series deep learning methods: (a) TC3-1; (b) TC3-11; (c) TC3-12; (d) TC3-17.
Figure 10. Predictive performance comparison between BiGRU and three canonical time-series deep learning methods: (a) TC3-1; (b) TC3-11; (c) TC3-12; (d) TC3-17.
Sustainability 17 07401 g010
Figure 11. Box plots comparing the residual distributions of the four models at each monitoring point: (a) TC3-1, (b) TC3-11, (c) TC3-12, and (d) TC3-17.
Figure 11. Box plots comparing the residual distributions of the four models at each monitoring point: (a) TC3-1, (b) TC3-11, (c) TC3-12, and (d) TC3-17.
Sustainability 17 07401 g011aSustainability 17 07401 g011b
Table 1. Information of Baise Water Control Project.
Table 1. Information of Baise Water Control Project.
ItemsValues
Drainage Area Proportion47.57%
Average Annual Flow263.0 m3/s
Annual Runoff8.29 billion m3
Normal Pool Level (NPL)228.0 m
Reservoir Area135 km2
Backwater Length108 km
Design Flood Level229.66 m
Check Flood Level231.27 m
Flood Control Restriction Level214.0 m
Dead Water Level203.0 m
Total Reservoir Capacity5.66 billion m3
Flood Control Capacity1.64 billion m3
Regulating Capacity2.62 billion m3
Dead Storage Capacity2.18 billion m3
Hydropower Station Capacity540 MW
Navigation Structure Design2 × 500-ton-class ships
Engineering CategoryClass I engineering structure
Table 2. Comparison of model performance for different rolling window sizes.
Table 2. Comparison of model performance for different rolling window sizes.
Window SizeMetrics2345710
TC3-1RMSE1.511.441.621.591.611.58
MAE1.050.981.181.131.161.14
R20.870.890.850.860.860.85
TC3-11RMSE2.091.542.162.074.213.21
MAE1.520.911.461.893.742.72
R20.850.930.840.850.450.67
TC3-12RMSE2.212.062.162.162.942.59
MAE1.721.431.611.522.291.84
R20.920.940.920.930.840.86
TC3-17RMSE1.841.761.811.841.981.97
MAE1.441.391.421.541.641.57
R20.770.810.790.720.690.71
Table 3. Optimal hyperparameters for the BiGRU model.
Table 3. Optimal hyperparameters for the BiGRU model.
HyperparameterOptimal Value
Units (Layer 1)160
Dropout Rate (Layer 1)0.3
Units (Layer 2)80
Dropout Rate (Layer 2)0.2
Learning Rate0.001
Table 4. Experimental environment and model hyperparameters.
Table 4. Experimental environment and model hyperparameters.
CategoryItemSpecification
HardwareCPUIntel(R) Core (TM) i7-10700 CPU @ 2.90 GHz
GPUNVIDIA GeForce GT 730
Memory (RAM)15.74 GB
Operating SystemWindows 10
SoftwarePython3.11.4
TensorFlow2.18.0
Scikit-learn1.6.1
Pandas2.2.3
Numpy2.0.2
Kerastuner1.0.5
HyperparametersModel Architecture2-Layer Bidirectional GRU
Units (Layer 1)160
Dropout Rate (Layer 1)0.3
Units (Layer 2)80
Dropout Rate (Layer 2)0.2
OptimizerAdam
Learning Rate0.001
Loss FunctionMean Squared Error (MSE)
Batch Size32
EpochsUp to 500, with Early Stopping (Patience = 20)
Table 5. Predictive performance comparison of BiGRU against three canonical time-series deep learning models.
Table 5. Predictive performance comparison of BiGRU against three canonical time-series deep learning models.
Monitoring PointMetricsLSTMBiLSTMGRUBiGRU
TC3-1RMSE2.212.311.651.44
MAE1.681.671.170.98
R20.730.700.850.89
TC3-11RMSE2.313.973.141.54
MAE1.693.182.690.91
R20.840.480.700.93
TC3-12RMSE3.063.992.072.06
MAE2.563.281.551.43
R20.860.740.930.94
TC3-17RMSE2.212.161.651.76
MAE1.831.711.301.39
R20.710.690.840.81
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ma, J.; Huang, X.; Wu, H.; Yan, K.; Liu, Y. Bidirectional Gated Recurrent Unit (BiGRU)-Based Model for Concrete Gravity Dam Displacement Prediction. Sustainability 2025, 17, 7401. https://doi.org/10.3390/su17167401

AMA Style

Ma J, Huang X, Wu H, Yan K, Liu Y. Bidirectional Gated Recurrent Unit (BiGRU)-Based Model for Concrete Gravity Dam Displacement Prediction. Sustainability. 2025; 17(16):7401. https://doi.org/10.3390/su17167401

Chicago/Turabian Style

Ma, Jianxin, Xiaobing Huang, Haoran Wu, Kang Yan, and Yong Liu. 2025. "Bidirectional Gated Recurrent Unit (BiGRU)-Based Model for Concrete Gravity Dam Displacement Prediction" Sustainability 17, no. 16: 7401. https://doi.org/10.3390/su17167401

APA Style

Ma, J., Huang, X., Wu, H., Yan, K., & Liu, Y. (2025). Bidirectional Gated Recurrent Unit (BiGRU)-Based Model for Concrete Gravity Dam Displacement Prediction. Sustainability, 17(16), 7401. https://doi.org/10.3390/su17167401

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop