1. Introduction
Hydraulic buildings like dams play a critical role in sustainable water resource management and hydropower generation, serving as key infrastructures for water storage, flood control, and energy supply in water-scarce and hazard-prone regions [
1,
2,
3,
4]. Hydraulic buildings are often built in regions with complex geological conditions, requiring advanced engineering techniques to ensure structural stability and longevity [
5,
6]. Given the immense scale and complexity of these structures, even small deformations can have significant impacts on the safety and operational efficiency of hydraulic buildings [
7]. Therefore, effective deformation monitoring and predictive maintenance of hydraulic structures are essential to prevent potential failures that could lead to catastrophic consequences, both economically and environmentally.
Statistical regression methods, such as multiple linear regression(MLR) and generalized linear models are widely used in deformation monitoring for hydraulic structures due to the simplicity and ability to fit relationships between monitoring variables and deformation [
8,
9,
10]. These methods rely on historical and real-time data collected from structural health monitoring systems, such as deformation, strain, temperature, rainfall, and water pressure, to establish predictive models that can estimate dam deformation behavior under different operational and environmental conditions [
11,
12]. However, statistical modeling methods rely on the assumption of linearity between the independent and dependent variables. In real-world applications, the relationship between the deformation-related variables is highly non-linear, leading to reduced prediction accuracy [
13]. Statistical models struggle to capture complex, high-dimensional interactions between deformation-related variables, but they may oversimplify these interactions, potentially overlooking critical deformation mechanisms.
With the rapid advancement of information technology, the research and application of artificial intelligence (AI) and deep learning (DL) techniques in dam deformation prediction have become increasingly prominent [
14,
15,
16]. Unlike traditional statistical approaches, DL-based methods can automatically extract and learn deep features from monitoring data without relying on predefined functional relationships, making them more suitable for addressing the complex, multi-factor coupling issues involved in dam deformation [
17]. For instance, Zhang et al. [
18] integrated densely connected convolutional networks and Long short-term memory (LSTM) networks to demonstrate superior accuracy and generalization in predicting the dynamic deformation of concrete gravity dams, achieving a correlation coefficient above 0.99. Xu et al. [
19] employed Max-Relevance and Min-Redundancy and Lasso for feature selection and Convolutional Neural Network(CNN)-LSTM for dam deformation and crack opening prediction, thus providing a robust framework for the safe monitoring of cracked concrete arch dams. Lu et al. [
20] proposed an Inception–ResNet–Gated Recurrent Unit (GRU) model that integrates feature extraction with long-term dependency learning, significantly improving dam deformation prediction accuracy for enhanced safety monitoring. Pan et al. [
5] developed a CNN-based spatiotemporal deformation field model incorporating temperature, water pressure, and constraint fields for arch dams without the need for inverse analysis or parameter adjustment. Based on the above analysis, it can be inferred that DL-based methods show potential values in capturing the temporal and spatial dependencies within dam monitoring data. However, traditional single-layer DL networks, such as LSTM, lack the depth and hierarchical learning capacity required to capture complex interactions influenced by water pressure, temperature variations, and structural stress, limiting the predictive performance in long-sequence, multi-variable contexts. Another major limitation of traditional LSTM-based models is their unidirectional nature, as they rely solely on past observations to predict future outcomes, overlooking the potential influence of future conditions on the current state [
21]. In addition, the predictive performance of LSTM architectures relies heavily on hyperparameter selection, including layer count, units per layer, dropout rates, learning rates, and time steps. Interactions among these hyperparameters greatly affect the convergence, generalization, and accuracy of predictive models [
22]. As model complexity increases, so does computational cost, leading to longer training times and a heightened risk of overfitting if hyperparameter tuning is not carefully managed. Traditional methods like grid or random search are often inefficient for navigating the vast hyperparameter space of complex LSTM models.
Based on the above analysis, this study proposes a novel DL-based deformation monitoring model for utilizing high-dimensional and long-sequence hydrostatic and thermal time series. First, an improved dam deformation monitoring model considering actual long-sequence prototypical hydrostatic and thermal data is developed, incorporating extensive thermometer data specifically for high arch dams. Next, a hybrid DL network architecture, featuring a Bi-Stacked Long Short-Term Memory (Bi-Stacked-LSTM) network, is constructed to capture the intricate temporal dynamics and complex interdependencies inherent in deformation-related hydrostatic and thermal sequences. Then, the Gaussian Process-based Surrogate Model Optimization (GP-SMO) algorithm is implemented to efficiently navigate the extensive hyperparameter space of DL techniques, optimizing parameters such as the number of LSTM units per layer, LSTM layer count, dense layer structures, dropout rates, and time steps. Multiple deformation monitoring points from a 300 m-high arch dam are analyzed. Comparative analysis is conducted using a range of advanced statistical and machine learning (ML) models, with predictive performance evaluated through diverse qualitative and quantitative metrics.
The key contributions of this study are as follows:
The proposed Bi-Stacked-LSTM model can capture complex nonlinear relationships between deformation and long-sequence hydrostatic and thermal time series in hydraulic structures by integrating multiple bidirectional LSTM layers, enabling it to effectively leverage intricate temporal dynamics and dependencies within the deformation data.
The GP-SMO algorithm can efficiently identify the optimal parameter configurations within a high-dimensional space, aligning the DL-based model with the unique temporal dependencies and nonlinear relationships in dam deformation monitoring data.
Through extensive quantitative and qualitative evaluations across various advanced ML models, the developed DL-based method demonstrates robust predictive power and reliability in dam deformation forecasting, validated from multiple perspectives in a series of comparative experiments.
The structure of this paper is as follows:
Section 2 presents the GP-SMO-based Bi-Stacked-LSTM deformation monitoring model for hydraulic structures, focusing on neuron configurations and hyperparameter optimization.
Section 3 covers dam monitoring data sources and preprocessing for training.
Section 4 discusses the training process, hyperparameter optimization, and model evaluation across multiple monitoring points. Finally,
Section 5 summarizes the contributions, advantages, limitations, and future research directions.
2. Proposed Method
2.1. Flowchart of the Developed Model
Figure 1 shows the flowchart of the proposed deformation prediction framework for high arch dams, comprising predictive model construction and hyperparameter optimization process of DL methods. In the predictive model stage, the Bi-Stacked-LSTM architecture is constructed and employed to capture the nonlinear and long-term dependencies within environmental monitoring data and dam deformation sequences. Environmental monitoring data (water level, temperatures, and rainfall) and historical deformation data are normalized and partitioned into training and testing sets, which are then fed into DL models, allowing the model to learn temporal patterns effectively. In the hyperparameter optimization process, the GP-SMO algorithm is utilized to optimize critical hyperparameters of DL models (e.g., the number of LSTM layers, neuron counts in each layer, dropout rate, and time steps). The developed surrogate model guides the sampling and updating of hyperparameters to enhance model performance iteratively, ensuring that the optimal parameters are achieved for accurate predictions.
2.2. Deformation Monitoring Model for Dam Structures
Dam deformation, a characteristic structural response, can be divided into recoverable components—influenced by water pressure and temperature—and irrecoverable components, which are affected by creep, alkali-aggregate reactions, and material aging [
23]. The loads and environmental impacts of dams in long-term service can be seen in
Figure 2.
In the statistical model, dam deformation
can be described using hydraulic component
, thermal component
, and time-varying component
. Specifically,
denotes the elastic deformation under hydraulic load,
denotes the recoverable deformation affected by temperatures, and
denotes the irreversible dam deformation caused by dam material aging. The details are as follows:
where for gravity dams,
; for arch dams,
. In this study, the object of this study is an arch dam, so the coefficient is 4.
The difference between the Hydraulic–Seasonal–Time (HST) and Hydraulic–Temperature–Time (HTT) models mainly exists in the interpretation and simulation of temperature-related variables [
24]. In the HST model, the temperature effect is mainly simulated by simple harmonics, while the HTT model is mainly simulated by a large number of thermometer monitoring data. The specific contents are as follows:
The time-varying component
of the deformation of a high arch dam can be expressed by the following formula:
Based on the HST and HTT deformation monitoring models, the deformation of high arch dams can be characterized as follows:
where
;
denotes the cumulative days from the initial monitoring date to the current one and represents the total number of thermometers embedded within the dam and its foundation;
represents the number of lag components; and
,
,
,
,
,
, and
denote the regression coefficients. The relevant models can be obtained from
Table 1.
2.3. Multi-Layer Bi-Stacked-LSTM Network
LSTM networks, a type of Recurrent Neural Network (RNN), are widely studied in time series forecasting for effectively addressing the vanishing and exploding gradient issues common in traditional RNNs.
Figure 3 depicts the architecture of LSTM architectures. It can be seen that LSTMs are equipped with memory cells and gating mechanisms (input, forget, and output gates) that control information flow, enabling the model to capture long-term dependencies in sequential data. These gates determine which information to retain or discard at each time step, allowing LSTMs to effectively model long-term patterns in the data.
At this moment, the forget gate is determined by
and
, and is expressed as follows:
where
is a function of
,
,
, and
are the weight and bias vector of the forget gate, respectively.
The input gate comprises two components: the gate signal and the candidate state, both generated by activation functions. These components are then combined through element-wise multiplication to update the cell state. The input gate controls the inflow of new information, selectively updating the cell state. The calculation formula is as follows:
where
is the cell state from the previous time step, whose value output is modulated by the activation function to control the degree of forgetting of the previous cell state;
is the input gate, the value output by the activation function, which controls the degree of retention of the candidate cell state
, output by the activation function.
where
is the gate value of the output gate, which is output by the activation function
;
is the weight matrix of the output gate;
is the hidden layer state at the previous moment;
is the input data at the current moment;
is the bias of the output gate; and
is the activation function. The output gate value, calculated by this expression, determines which part of the current cell state is passed to the external state. This controlled portion of the cell state is then output as the final result.
In the context of high arch dam deformation, which is influenced by multi-dimensional time series data (e.g., water levels, rainfall, and temperature), a single LSTM model often falls short in capturing the intricate interactions between these environmental factors and the dam’s deformation behavior. Furthermore, single LSTMs struggle to effectively model both short-term and long-term dependencies within the data, such as the immediate impact of rainfall versus the seasonal effects of temperature fluctuations.
To address these challenges, this study designs a Bi-Stacked-LSTM network architecture specifically for predicting multi-dimensional, long-sequence time series related to high arch dam deformation.
Figure 4 depicts the schematic diagram of the developed Bi-Stacked-LSTM model. It can be seen that the developed model could better capture the intricate, multi-scale dependencies in the deformation-related time series, allowing for more accurate predictions of dam deformation under the influence of diverse and interdependent environmental factors. The Bi-Stacked-LSTM model processes dam deformation-related time series data in both forward and backward directions, capturing insights from past and future trends. This bi-directional and stacked-layer architecture enhances the model’s ability to learn both short-term and long-term patterns in high arch dam deformation, effectively addressing the complex interactions between environmental factors and deformation monitoring data for more accurate predictions.
2.4. The Improved GP-SMO Algorithm
In the context of dam deformation analysis, environmental factors (e.g., water levels, temperature, rainfall, and time-varying factors) have complex, nonlinear relationships with structural deformation. Traditional physics-based models often struggle to capture these relationships accurately. A Gaussian Process (GP)-based surrogate model, grounded in Bayesian theory, can effectively represent the statistical dependency between observed environmental factors and dam deformation, offering a predictive approach for assessing dam deformation performance under varying conditions.
A Gaussian Process is a non-parametric Bayesian model that assumes any finite set of function observations follows a multivariate normal distribution. For a given target function
, where
represents a d-dimensional input vector of environmental variables, GP assumes the following:
where
is the mean function, often set to zero for simplicity, and
is the covariance (or kernel) function, which defines the similarity between two input points
and
. Commonly used kernels, such as the Radial Basis Function (RBF) or Matern kernel, capture dependencies among environmental variables, which are crucial for modeling their impact on deformation.
Given a set of
observed dam deformation monitoring points
, where
represents the observed environmental factors and
the corresponding dam deformation measurements, Bayesian inference enables us to make predictions at new input locations. Specifically, for an unobserved point
, the posterior distribution of the predicted deformation
given all observations is as follows:
where
is the covariance vector between the new point
and the observed points;
denotes the covariance matrix among the observed points; and
is the noise variance associated with observation noise.
To enhance the predictive accuracy for dam deformation under new environmental conditions, Bayesian optimization is often applied in tandem with the GP surrogate model. Bayesian optimization uses an acquisition function to determine where to sample next. These acquisition functions balance exploration and exploitation, thereby achieving efficient global optimization even with limited sampling.
Figure 5 illustrates the iterative process of the GP-SMO optimization algorithm, often applied in scenarios such as dam deformation analysis, where experimental costs are high and the objective function is complex and nonlinear. The process consists of two main phases: Inference and Planning, which operate in a loop to iteratively refine the model and optimize the target objective. This approach combines Bayesian inference with strategic experimental planning to enhance model performance in predicting target outcomes with limited deformation-related data. The specific calculation steps are as follows:
Step 1: Initial Experiments: The process begins with a set of initial experiments , where represents the environmental inputs, and yiy_iyi represents the measured responses, such as dam deformation.
Step 2: Updating the Surrogate Model: The surrogate model is then updated based on the initial data, mapping xxx to predictive mean and variance . These statistical outputs represent the expected behavior of the response and the uncertainty around it, respectively, under different environmental conditions.
Step 3: Planning with Acquisition Functions: Using the posterior predictions and , an acquisition function is computed. Common acquisition functions can help to balance exploration (searching in uncertain regions) and exploitation (focusing on promising regions) in the search for optimal solutions. The next experimental point is selected by maximizing .
Step 4: Iterative Experimental Updates: The selected experiment is added to the dam environmental monitoring dataset, and its outcome is recorded. The surrogate model is then updated with this new data, thereby improving its accuracy and refining its predictions iteratively.
Step 5: Objective Optimization and Benchmarking: This cycle continues, iteratively updating the surrogate model and refining the acquisition function until the optimal input is found. The process concludes with benchmarking using evaluation metrics to assess effectiveness and accuracy.
4. Experimental Results Analysis
The DL-based dam deformation prediction model and comparative methods were implemented on a graphics workstation featuring an AMD Ryzen 9 5900X CPU, NVIDIA RTX 4080 GPU, 64 GB RAM, and 2 TB SSD, running TensorFlow on an Ubuntu Linux OS optimized for GPU-accelerated tasks and high-level computations. Multiple evaluation metrics, including Mean Absolute Error (MAE), symmetric Mean Absolute Percentage Error (sMAPE), Root Mean Square Error (RMSE), and the Coefficient of Determination (R_squared), were employed to comprehensively assess and compare the performance of dam deformation predictive methods.
4.1. The Model Hyperparameter Optimization Process
Figure 13 demonstrates the changes in loss values and R_squared evaluation metrics during the GP-SMO parameter tuning process. It can be seen from
Figure 13a that the parameter optimization process has achieved a significant loss reduction, particularly after around 90 epochs, where the minimum value of 0.0035 is observed. This suggests that the developed GP-SMO algorithm can accurately identify a set of hyperparameters that significantly improve the model’s predictive ability to generalize to the training data. The fluctuation in loss before reaching the optimal point indicates that the model was undergoing a phase of exploration in the parameter space, leading to a better configuration. It can be seen from
Figure 13b that the highest R_squared value of 0.9548 occurs at epoch 90, coinciding with the point where the loss reaches its minimum. The R-squared score reflects the proportion of variance explained by the model, and a value close to 1 signifies an excellent fit of the model to the data.
Figure 14 depicts the variation of the loss function during the training process of the proposed Bi-Stacked-LSTM model using the obtained optimal hyperparameters. It can be seen that the loss starts at a relatively high value of approximately 0.45, initially reflecting the initial prediction error of the model when the training begins. However, the loss quickly decreases within the first 50 epochs, indicating that the model rapidly learns from the data and reduces the prediction error. As the epochs progress beyond 50, the loss continues to decrease but at a much slower rate, suggesting that the model is converging and refining its predictions with each additional epoch. By around epoch 100, the loss stabilizes and approaches a very low value close to zero, remaining nearly constant for the remainder of the training period up to 500 epochs.The relevant optimized deep learning model hyperparameters are shown as
Table 3.
4.2. Model Ablation Experiment
Referring to the optimal hyperparameters obtained based on the GP-SMO algorithm in the previous section, the GP-SMO-Bi-Stacked-LSTM model for high arch dam deformation prediction was constructed. To further verify the effectiveness of the various components of the GP-SMO-Bi-Stacked-LSTM proposed in this study for the accuracy of dam deformation prediction, five comparative algorithms, namely GP-SMO-Stacked-LSTM, GP-SMO-Bi-LSTM, Stacked-LSTM, GP-SMO-LSTM model, and single LSTM model, were introduced to conduct ablation experiments on each component.
Table 4 shows the ablation test evaluation results of each component of the proposed method on the test set. It should be noted that the evaluation results are based on the average apple index results of the prediction results of three typical measurement points. As can be seen from
Table 4, the GP-SMO-Bi-Stacked-LSTM neural network model learns long-term dependencies and trend change laws based on dam deformation-related data for sample training and has better prediction ability on the deformation prediction in the test sets. The prediction accuracy is significantly higher than other algorithms. This further verifies the advantages of the GP-SMO algorithm in optimizing the hyperparameters of the Bi-Stacked-LSTM model. In other words, the GP-SMO optimization algorithm effectively explores the parameter architecture and optimal hyperparameters of the constructed Bi-Stacked-LSTM model and thus obtains a more reasonable parameter and architecture model of the Bi-Stacked-LSTM.
In contrast, the prediction effect of other models may be limited by insufficient hyperparameter tuning or the limitations of the model itself. The single LSTM model has problems of insufficient model fitting ability and inappropriate eigenvalues because it fails to fully utilize the causal feature information of multiple factors, resulting in lower dam deformation prediction accuracy than the proposed method. The prediction accuracy of the GP-SMO-LSTM model under various indicator evaluations is higher than that of the single LSTM model. This is mainly because the GP-SMO-LSTM model better balances the complexity and prediction ability of the model, so its performance on the training set is more accurate. However, since the GP-SMO-LSTM model still uses a single LSTM hidden layer, the feature extraction ability is insufficient, resulting in limited generalization performance on dam deformation prediction. The prediction accuracy of the GP-SMO-Stacked-LSTM and GP-SMO-Bi-LSTM models is higher than that of the GP-SMO-LSTM model, which shows that under the premise of parameter optimization, the bidirectional transmission mechanism and stacking strategy can improve the feature extraction ability of the network model to a certain extent, thereby improving the model dam deformation prediction ability.
4.3. Performance Evaluation of Comparative Methods
To validate the effectiveness of the proposed DL-based deformation prediction model for high arch dams, a series of statistical and ML-based comparative methods were introduced. These methods included HST-based multiple linear analysis (HST-MLR), Light Gradient Boosting Machine (LGBM), Random Forest (RF), Multi-Layer Perceptron (MLP), Support Vector Regression (SVR), Long Short-Term Memory (LSTM), and Gradient Boosting Machine (GBM). To ensure a fair comparison, all comparative methods and the proposed DL-based model were trained and evaluated on the same dataset under identical environmental conditions. This approach guarantees that the observed differences in model performance are attributable solely to the model architectures and algorithms rather than variations in data or experimental setup. The principles of the relevant methods are introduced as follows:
- (1)
HST-MLR model. The HST-based MLR model is widely used in dam safety monitoring, which analyzes the relationship between multiple independent variables and a single dependent variable. It assumes a linear relationship, where each predictor has a unique coefficient indicating its contribution to the outcome [
25].
- (2)
LGBM model. LGBM is a high-performance gradient-boosting framework optimized for large datasets and high-dimensional data. Unlike traditional boosting, LGBM grows trees leaf-wise, thus improving computational speed and memory efficiency. This makes it suitable for applications requiring rapid, accurate predictions with reduced computational cost [
26].
- (3)
RF model. An ensemble learning method, RF constructs multiple decision trees and aggregates their outputs, resulting in robust generalization and reduced overfitting. It is highly effective for handling feature interactions and noisy data, making it a popular choice for regression tasks [
27].
- (4)
MLP model. MLP is a type of feedforward neural network consisting of multiple layers of neurons, enabling it to capture complex, non-linear relationships within the data. It performs well in both regression tasks but requires significant computational resources and training data for optimal performance [
28].
- (5)
SVR model. An extension of Support Vector Machines, SVR is designed for regression tasks and minimizes an epsilon-insensitive loss function. This method is particularly effective in handling high-dimensional data and ensuring robust performance in the presence of noise [
29].
- (6)
LSTM model. LSTM, a specialized recurrent neural network architecture, effectively captures sequential dependencies by mitigating the vanishing gradient problem. It uses memory cells and gating mechanisms, allowing it to retain long-term information, making it ideal for time-series prediction [
30].
- (7)
GBM model. The GBM model ensemble method builds models sequentially, with each new model aiming to correct errors made by previous ones using gradient descent on a specified loss function [
31].
Table 5 shows the hyperparameter range settings and optimization results of comparison methods. It can be inferred that corresponding hyperparameters have been added to this manuscript to list the specific values or search ranges used for each key hyperparameter (e.g., number of estimators, learning rate, maximum depth, number of hidden units, etc.), along with the optimization method (e.g., grid search, default setting, or manual tuning based on validation performance).
Three representative monitoring points (TCN08 to TCN10) were selected as study targets to illustrate distinct deformation patterns that are directly indicative of the dam’s structural response under varying hydrostatic and thermal influences.
Figure 15 shows the prediction and residual analysis for dam deformation at three monitoring points from TCN08 to TCN10. It consists of six subplots, where subplots (a), (b), and (c) show the predicted deformation results, and subplots (d), (e), and (f) display the residuals associated with these predictions.The evaluation index calculation results of the proposed and compared methods are shown in
Table 6. It can be inferred from
Figure 15a–c that the proposed GP-SMO-based Bi-Stacked-LSTM method demonstrates superior performance in capturing the temporal trends of dam deformation in the prediction results for monitoring points TCN08, TCN09, and TCN10. Compared to other benchmark comparative methods such as GBM, HST-MLR, LGBM, LSTM, MLP, RF, and SVR, the proposed method aligns more closely with the observed values, indicating higher predictive accuracy.
Figure 15d–f demonstrates the residual plots, which indicate the error distribution for each method’s predictions relative to the actual deformation values. In terms of residuals, the proposed method consistently shows smaller and less dispersed residuals around the zero line across all monitoring points. This suggests fewer and smaller prediction errors, as well as improved stability relative to other methods. The lower variability in the residuals further supports the robustness and reliability of the proposed method in accurately predicting dam deformation. Consequently, the proposed method offers a more reliable tool for predicting dam deformation, providing critical insights for the safety monitoring of high arch dams and decision-making processes.
4.4. Graphical Visualization Evaluation Comparison
Figure 16 presents radar charts comparing the performance of different ML-based models across R-squared, RMSE, and sMAPE metrics at three monitoring points. In terms of R-squared (subplots a, d, g), which reflects the proportion of variance explained by each model, the proposed method generally covers the largest area, indicating superior performance in capturing variance in dam deformation data. For RMSE (Subplots b, e, h), a metric assessing the standard deviation of prediction errors, the proposed method consistently shows lower values, suggesting higher predictive accuracy and reduced error spread. Lastly, for sMAPE (Subplots c, f, i), which focuses on percentage errors, the proposed method achieves the lowest values across monitoring points, further demonstrating its advantage in prediction accuracy over other models.
In summary, the radar charts provide a comparative view of model performance across multiple metrics and monitoring points. The proposed method consistently shows superior performance in terms of higher R-squared values and lower RMSE and sMAPE values, indicating its effectiveness in predicting dam deformation more accurately than the other models tested.
Figure 17 demonstrates the residual box plots of the proposed and other benchmark methods at monitoring points TCN08, TCN09, and TCN10, respectively. It can be inferred that the proposed method consistently exhibits lower median residuals and a narrower interquartile range across all monitoring points, highlighting its accuracy and stability in dam deformation prediction. Additionally, the proposed method has fewer extreme outliers, particularly when compared to methods such as HST-MLR, RF, and LSTM, which display higher variability and a wider spread of residuals. This suggests that the proposed method not only reduces prediction error but also enhances robustness by minimizing residual dispersion, making it a more reliable choice for high-precision deformation monitoring in dam safety assessments.
Figure 18 presents the long-term deformation prediction results of the proposed method at three monitoring points (TCN06, TCN07, and TCN08) from 2008 to 2017. The predicted values (red curves) exhibit strong agreement with the measured deformation data (blue curves), successfully capturing both seasonal fluctuations and long-term trends. The proposed model demonstrates consistent performance across different deformation magnitudes and monitoring locations, indicating its robustness and reliability in practical dam deformation structural health monitoring applications.
5. Conclusions
The monitoring of deformation in hydraulic structures is critical for maintaining structural integrity and ensuring safe, continuous operation under varying hydraulic conditions. These structures are constantly subjected to environmental stressors, particularly long-term fluctuations in hydrostatic pressure and temperature. Thus, accurate prediction of deformation patterns becomes essential for proactive maintenance and water-related risk mitigation. However, traditional statistical regression and shallow ML-based approaches are often constrained by their linear modeling assumptions, dependency on hand-crafted deformation features, and limited capacity to capture complex temporal dynamics driven by environmental variability.
To overcome these limitations, this study proposes a Bidirectional Stacked Long Short-Term Memory (Bi-Stacked-LSTM) network, specifically designed to model intricate temporal dependencies and directional interactions embedded in multivariate deformation time series induced by hydrostatic and thermal conditions. To further enhance model performance, hyperparameters such as the number of LSTM layers, neurons per layer, dropout rate, and time steps are efficiently optimized using the Gaussian Process-based Surrogate Model Optimization (GP-SMO) algorithm. The framework is validated using long-term monitoring data collected from multiple deformation measurement points in the crown beam region of high arch dams. The main conclusions drawn from this study are as follows:
- (1)
The proposed Bi-Stacked-LSTM model effectively captures the complex nonlinear relationships between deformation and deformation-related environmental variables (e.g., long-sequence hydrostatic and thermal time series) in hydraulic structures by integrating multiple bidirectional LSTM layers. This architecture allows the developed model to leverage intricate temporal dynamics and dependencies within the deformation data, resulting in enhanced predictive accuracy.
- (2)
The GP-SMO algorithm efficiently identifies optimal parameter configurations within a high-dimensional space, ensuring that the deep learning-based model aligns with the unique temporal dependencies and nonlinear relationships present in dam deformation monitoring data. This optimization process not only enhances model performance but also enables more accurate predictions, ultimately contributing to improved structural health monitoring and risk management strategies for high arch dams.
- (3)
Through extensive quantitative and qualitative evaluations of various advanced ML-based models, the developed DL-based method demonstrates robust predictive power and reliability in forecasting deformation in hydraulic structures. This effectiveness is validated from multiple perspectives through a series of comparative experiments, highlighting its superiority in accurately capturing the complexities of deformation patterns.
However, this research is not without its limitations. First, while the developed GP-SMO-based Bi-Stacked-LSTM model shows improved performance, its complexity may pose challenges in terms of interpretability. Stakeholders may find it difficult to derive actionable insights from the model’s predictions due to the intricate nature of its architecture. Additionally, the reliance on historical data to train the dam deformation model may lead to overfitting, particularly in scenarios with limited datasets. Future work could explore strategies to mitigate this risk, such as employing regularization techniques or expanding the dam monitoring dataset through synthetic data generation. Moreover, the prediction model of deformation is contingent upon the quality and granularity of the input data. In practice, variations in data collection methods, sensor accuracy, and environmental factors can significantly impact predictive performance. Thus, further research is needed to evaluate the model’s robustness across diverse datasets and under varying conditions. In addition, the authors plan to further expand this analysis in future work by incorporating all available dam deformation monitoring points to investigate spatiotemporal dependencies and to enhance the model’s generalization across the entire dam structure.
Lastly, while the focus of this study was on deformation prediction, the proposed framework could be adapted for other hydraulic structural response variables, such as seepage, crack opening, etc. Future studies may investigate its applicability to different types of dams like concrete face rockfill dams and gravity dams, further extending the contributions of this research. In conclusion, while the developed DL-based model offers significant advancements in dam deformation predictive accuracy, addressing the highlighted limitations will be essential for its successful implementation in real dam safety management applications.