Next Article in Journal
GCMark: Robust Image Watermarking with Gated Feature Selection and Cover-Guided Expansion
Previous Article in Journal
Weighted Lp Estimates for Multiple Generalized Marcinkiewicz Functions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Streamflow Prediction of Spatio-Temporal Graph Neural Network with Feature Enhancement Fusion

1
School of Information Engineering, Nanjing Xiaozhuang University, Nanjing 211171, China
2
School of Cyber Security, Tianjin University, Tianjin 300072, China
3
Nari Group Corporation (State Grid Electric Power Research Institute), Nanjing NARI Information and Communication Technology Co., Ltd., Nanjing 211171, China
4
School of Computer Science and Technology, Anhui University of Technology, Ma’anshan 243032, China
5
College of Computer Science and Software Engineering, Hohai University, Nanjing 210098, China
*
Author to whom correspondence should be addressed.
Symmetry 2026, 18(2), 240; https://doi.org/10.3390/sym18020240
Submission received: 18 December 2025 / Revised: 11 January 2026 / Accepted: 26 January 2026 / Published: 29 January 2026
(This article belongs to the Section Computer)

Abstract

Despite the promise of graph neural networks (GNNs) in hydrological forecasting, existing approaches face critical limitations in capturing dynamic spatiotemporal correlations and integrating physical interpretability. To bridge this gap, we propose a spatial-temporal graph neural network (ST-GNN) that addresses these challenges through three key innovations: dynamic graph construction for adaptive spatial correlation learning, a physically-informed feature enhancement layer for soil moisture and evaporation integration, and a hybrid Graph-LSTM module for synergistic spatiotemporal dependency modeling. The temporal and spatial modules of the spatio-temporal graph neural network exhibit a structural symmetry, which enhances the model’s representational capability. By integrating these components, the model effectively represents rainfall-runoff processes. Experimental results across four Chinese watersheds demonstrate ST-GNN’s superior performance, particularly in semi-arid regions where prediction accuracy shows significant improvement. Compared to the best-performing baseline model (ST-GCN), our ST-GNN achieved an average reduction in root mean square error (RMSE) of 6.5% and an average improvement in the coefficient of determination (R2) of 1.8% across 1–8 h forecast lead times. Notably, in the semi-arid Pingyao watershed, the improvements reached 13.3% in RMSE reduction and 2.5% in R2 enhancement. The model incorporates watershed physical characteristics through a feature fusion layer while employing an adaptive mechanism to capture spatiotemporal dependencies, enabling robust watershed-scale forecasting across diverse hydrological conditions.

1. Introduction

Hydrological forecasting plays a critical role in water resource management, flood mitigation, and sustainable development [1,2]. Traditional approaches, including physical-based models and conventional machine learning methods (e.g., ANN [3,4,5], LSTM [6,7,8], and CNN [9,10,11,12]), often face challenges in capturing complex spatial dependencies and nonlinear dynamics inherent in hydrological systems [13,14,15].
In hydrological forecasting, the intrinsic graph structure of monitoring networks presents a natural alignment with graph-based modeling approaches. Rainfall stations and streamflow gauges constitute nodes in a network, while their spatial arrangements and temporal interactions define edge relationships. Graph neural networks (GNNs [16]) have emerged as a principled framework for such problems due to their capacity to explicitly model spatial dependencies through graph convolutions and to integrate temporal dynamics when combined with recurrent architectures. Recent advances in dynamic graph neural networks [17,18,19,20,21,22,23] further enhance this capability by enabling adaptive learning of time-varying relational structures, which is particularly relevant for capturing the evolving spatial correlations in hydrological systems. These characteristics address key challenges in rainfall-runoff modeling, including the tracking of moving rainfall centers and the capture of long-term historical dependencies. Moreover, graph-based representations offer flexibility in incorporating heterogeneous data types, such as physical distance metrics, statistical correlations between stations, and domain-specific variables like soil moisture and evaporation. Given these analytical advantages, the following review examines GNN-based methodologies that are particularly relevant to advancing streamflow prediction capabilities.
Liu et al. [24] proposed a directed graph deep neural network (DGDNN) that integrates spatial information capture (via convolutional layers) and feature aggregation (via MLPs) to simulate rainfall-runoff and confluence processes. Applied to the Yangtze River basin, DGDNN outperformed ANN, LSTM, GRU, and CNN, achieving higher accuracy and effectively quantifying uncertainty using hidden Markov regression (HMR). Kazadi et al. [25] developed FloodGNN, a GNN-based framework that processes vectorized water velocity features using geometric vector perceptrons (GVPs). Tested on Hurricane Harvey simulations, FloodGNN outperformed RNN baselines by jointly modeling spatiotemporal dynamics and preserving directional velocity information. The model demonstrated robustness in predicting flood extents and depths, even in high-flow regions, highlighting GNNs’ ability to handle vector-based hydrological features. Bai and Tahmasebi [26] employed Graph WaveNet, combining adaptive graph convolution with temporal dilated convolutions, to predict groundwater levels (GWL) across 41 wells in British Columbia. The model achieved superior performance by learning spatial dependencies through a self-adaptive adjacency matrix, even without prior knowledge of well connectivity. The learned adjacency matrix provided interpretable insights into inter-well influences, such as strong dependencies between geographically proximate wells.
Recent advancements have further refined GNN applications in hydrology with enhanced physical integration and architectural innovations. Taghizadeh et al. [27] developed interpretable physics-informed GNNs for flood forecasting, incorporating mass conservation constraints to improve physical consistency. Similarly, Ashraf et al. [28] applied physics-informed GNNs to water distribution systems, demonstrating improved generalization under data-scarce conditions. For runoff forecasting, Yang et al. [29] proposed a graph attention network with local-global-temporal attention mechanisms, effectively capturing multi-scale spatiotemporal dependencies. Li et al. [30] introduced an attention-based model for rainfall-runoff modeling, highlighting the importance of adaptive feature weighting. Beyond GNNs, transformer-based architectures are gaining traction: Yuan et al. [31] enhanced runoff prediction using causal lag-aware attention and multi-scale fusion in transformer models, while Sheng et al. [32] developed a residual temporal convolutional network with dual attention for interpretable multilead-time forecasting. Most recently, Zhou [33] proposed a multi-scale dynamic spatiotemporal graph attention network for karst spring discharge forecasting, emphasizing dynamic graph construction for time-varying hydrological connectivity.
These studies underscore GNNs’ versatility in hydrological applications, from simulating directed hydrological processes to handling vector features and unstructured spatial data. By integrating physical principles with data-driven learning, GNNs bridge the gap between traditional black-box models and physically interpretable frameworks, offering scalable solutions for multi-step, multi-site hydrological forecasting.
While the aforementioned studies demonstrate the promising potential of GNNs in hydrological forecasting, recent advances (2024–2025) have introduced sophisticated architectures that partially address some limitations, yet critical challenges remain for real-world streamflow prediction. For instance, Liu et al.’s DGDNN [24] effectively captures spatial rainfall-runoff processes through static graph constructions, yet its reliance on predefined adjacency matrices struggles with dynamic spatiotemporal correlations caused by migrating rainfall centers. Similarly, FloodGNN’s (Kazadi et al., [25]) geometric vector perceptron architecture excels in preserving directional flow velocity information but lacks mechanisms to model the nonlinear interactions between multi-scale historical hydrological features (e.g., rainfall from days/weeks prior) that fundamentally govern streamflow generation. Furthermore, while Graph WaveNet’s (Bai & Tahmasebi, [26]) self-adaptive adjacency learning shows interpretability benefits for well connectivity, its framework remains constrained by three key challenges.
Recent physics-informed approaches (Taghizadeh et al., [27], Ashraf et al., [28]) have improved model generalization through physical constraints but often employ static or simplified graph structures that cannot capture the time-varying nature of watershed connectivity. Attention-based models (Yang et al., [29], Li et al., [30]) enhance feature weighting but typically focus on either spatial or temporal attention separately, lacking integrated spatiotemporal attention mechanisms. Transformer-based architectures (Yuan et al., [31], Sheng et al., [32]) offer superior sequence modeling but often neglect explicit graph-based spatial representations crucial for watershed networks. Notably, Zhou’s [33] multi-scale dynamic spatiotemporal graph attention network represents progress in dynamic graph construction, yet its application remains limited to karst springs rather than general watershed streamflow prediction.
Collectively, these approaches face three persistent challenges inherent in medium-small watershed streamflow forecasting: (1) The need to dynamically represent spatial correlations between rainfall stations experiencing time-varying precipitation patterns, (2) The requirement to model dependencies across hydrological variables with heterogeneous temporal resolutions (e.g., hourly rainfall/streamflow vs. daily soil moisture/evaporation), necessitating specialized approaches for multi-resolution data fusion, and (3) The necessity to adaptively incorporate heterogeneous watershed characteristics (soil moisture, evaporation) and domain-specific knowledge within a unified architecture.
To bridge these gaps, we propose a novel spatial-temporal graph neural network (ST-GNN) framework that simultaneously addresses the dual challenges of complex spatiotemporal dynamics and model interpretability in streamflow prediction. Our architecture introduces three innovative components that collectively advance the state of the art:
  • We develop a dynamic graph construction mechanism that autonomously identifies and encodes spatial correlations between historical rainfall and flow data through adaptive graph convolution operations.
  • We present a physically informed feature enhancement layer that incorporates soil moisture and evaporation processes into the temporal feature propagation, enabling context-aware weighting of historical inputs.
  • We introduce a hybrid Graph-LSTM module that synergizes global spatial context learning with local temporal dependency modeling, effectively capturing both basin-scale hydrological interactions and station-specific temporal patterns.
By integrating these components, our framework not only achieves superior predictive accuracy but also provides actionable insights into the underlying spatiotemporal mechanisms governing streamflow generation.

2. Conceptual Framework

This section elaborates on the proposed ST-GNN prediction model, including its architectural framework, functional components, and input–output specifications for medium-small watershed streamflow forecasting.

2.1. ST-GNN Architectural Framework

The architecture of our ST-GNN is designed under the general paradigm of spatio-temporal deep learning for structured data, which integrates modules for spatial dependency modeling, feature fusion, and sequential processing [34]. Figure 1 illustrates the comprehensive framework of our ST-GNN, which consists of three core components: (1) multi-source input modules, (2) hierarchical processing modules, and (3) output prediction module. The processing modules integrate three specialized components: a graph convolutional network (GCN) for spatial dependency modeling, a feature-enhanced fusion layer for heterogeneous data integration, and a Graph-LSTM module for temporal dynamics capture.
The key components of the ST-GNN architecture, their functions, and input-output specifications are systematically summarized in Table 1. This table provides a concise overview of each module’s role within the processing pipeline, detailing the transformations applied to the data dimensions (where N is the number of stations, F the input features, D the GCN hidden dimension, D h the LSTM hidden dimension, T the input sequence length, and τ the forecast lead time). It serves as a reference for understanding the data flow and modular design of the proposed framework.

2.2. Multi-Source Input Modules

Effective streamflow forecasting requires the integration of multiple data sources that represent different aspects of the hydrological cycle [35]. The model incorporates three distinct input modalities to address key hydrological factors:
(1) History streamflow and rainfall data X: Temporal patterns of streamflow and rainfall observations, serving as primary predictors for short-term discharge forecasting.
(2) Watershed state indicators X state : Real-time measurements of soil moisture content and evaporation rates, which critically modulate rainfall-runoff transformation efficiency. Notably, low soil saturation or high evaporation significantly delays rainfall contribution to river discharge.
(3) Future rainfall trends X future : Anticipated precipitation trends that influence longer-term (1–8 h) discharge patterns at the watershed outlet.
The model output generates 1–8 h ahead forecasts of watershed outlet discharge.

2.2.1. History Streamflow and Rainfall Data

The spatial structure of the watershed is explicitly represented as a graph G = ( V , E ) , where each monitoring station (rainfall or streamflow gauge) constitutes a node v i V . Each node is associated with dynamic attributes (e.g., observed rainfall/streamflow time series) as well as static geographical features (e.g., latitude, longitude, and elevation) that define its position within the basin. As depicted in Figure 2, the spatial relationships—i.e., the edges E—between these nodes are encoded through two complementary graph construction methods [36]:
  • Distance graph
To represent the weight between stations in the distance graph, we use the reciprocal of the distance. The formalized distance graph is denoted as G D = V , A D , where V is the set of nodes (i.e., monitoring stations) and A D is the adjacency matrix that stores the reciprocal of the spatial channel distance as the edge weight.
A D i , j = 1 d i s ( r i , r j ) .
  • Correlation graph
To construct a dynamic, intra-basin correlation graph, we calculate the Pearson correlation coefficient between each pair of stations using a sliding window of historical monitoring data (e.g., the past 24 h). This captures the temporally evolving statistical dependencies within the watershed. The correlation at time t is used as the edge weight, defining a time-varying adjacency matrix A T ( t ) . This mechanism allows the model to adaptively adjust the strength of connections between stations based on their recent co-variation, reflecting how spatial hydrological connectivity can change during rainfall events.
The formalized correlation graph at time t is expressed as G T ( t ) = ( V , A T ( t ) ) . Here, V represents the set of nodes (i.e., monitoring stations), and A T is the adjacency matrix that stores the correlation between each pair of stations as the edge weight.
A T i , j = max 0 , ρ r i , r j ,
Here, ρ represents the Pearson correlation coefficient [37], whose output range is −1 to +1. A negative value indicates negative correlation, and a positive value indicates positive correlation.
The final adjacency matrix A ( t ) used for graph convolution at time step t is a normalized fusion of the static distance graph and the correlation graph: A ( t ) = Normalize A D + λ · A T ( t ) , where λ is a learnable parameter that balances the influence of physical proximity and statistical similarity, and Normalize ( · ) typically denotes row normalization to create a probability transition matrix.

2.2.2. Watershed State Indicators

Soil moisture and evaporation are recognized as pivotal state variables that govern the partitioning of rainfall into runoff and infiltration [38,39]. The initial state input comprises evaporation and soil moisture values. As these variables are typically recorded at daily intervals in operational monitoring networks, we apply linear temporal interpolation to generate hourly estimates that align with the rainfall and streamflow data timestamps. Before feeding into the model, the concatenated soil moisture and evaporation values are normalized via min–max scaling to ensure numerical stability and compatibility with other input features. The formalization is as follows:
X s t a t e = N o r m a l i z e C o n c a t X e v a , X s o i l ,
Here, X e v a represents the input for evaporation, and X s o i l represents the input for soil moisture content. C o n c a t [ · ] denotes channel-wise concatenation, and N o r m a l i z e ( · ) denotes min–max scaling.
The runoff generation mechanism of a natural watershed is a complicated process, and we can understand the role of these state variables by considering the complete runoff generation process. When the soil moisture content is below a certain threshold, most rainfall infiltrates to replenish the soil moisture; when the soil becomes saturated, subsequent rainfall predominantly produces runoff. The current evaporation state of the watershed is another influencing factor that is positively related to soil water depletion. Therefore, soil water content and evaporation state can significantly modulate the contribution of rainfall to streamflow, justifying their inclusion as dynamic state inputs to the model.

2.2.3. Future Rainfall Trends

Incorporating future rainfall information, often derived from numerical weather prediction models or satellite estimates, has been shown to improve streamflow forecast skill, especially for lead times beyond the concentration time of a watershed [8]. The third module of the model input is the future rainfall input. In recent years, high-resolution satellite remote sensing rainfall products have provided key input data for global and regional meteorological and hydrological research, especially in real-time monitoring and early warning of natural disasters caused by extreme rainfall. For small- and medium-sized watersheds, the future rainfall data can be calculated as the average of the monitoring data from each rainfall station. Therefore, the future rainfall input is defined as follows:
X f u t u r e = i = 1 N P i N ,
where N is the number of rainfall stations, and P i is the rainfall value of the i-th rainfall station.

2.3. Hierarchical Processing Modules

2.3.1. Spatial GCN Module

A spatial model that performs graph convolution locally at each node can easily share weights between different positions and structures. However, the spectrum-based model is limited to working on undirected graphs. Since the Laplacian matrix on the rainfall-runoff directed graph we constructed is not clearly defined, we use the space-based model to handle multi-source inputs more flexibly.
The hidden layer representation matrix of GCN [40] is defined as follows:
H k = f W k P k X ,
Here, f · is the activation function, and the probability transition matrix P is calculated by P = D 1 A . H k maintains the same dimension as the input X, and H 1 , H 2 , , H K are merged as the model output [41].
Since the diffusion process of a stationary distribution is performed by the sum of the power series of the probability transition matrix, the processing method of our reference [42] sums the output in each diffusion step, which is defined as follows:
H = k = 0 K f P k X W k ,
here, W k R D × F .

2.3.2. Feature Enhancement Fusion Module

We propose a novel feature enhancement fusion module to address the varying impact of hydrological conditions on streamflow generation. Different factors in a watershed can cause varying changes in streamflow. For instance, as depicted in Figure 3a, rainfall during the dry season may not cause significant changes in streamflow, while in the rainy season, high soil water content and rainfall may result in obvious streamflow responses (Figure 3b). Even with similar rainfall intensity and duration, the resulting flood discharges can be entirely different. To address this, we propose a feature enhancement fusion module in this model.
As shown in Figure 4, the feature enhancement fusion module’s input includes two parts: the output O 1 from the spatial GCN module and O 2 after pretreatment of evaporation and soil moisture. Inside the feature enhancement fusion module, we’ve set up a t a n h excitation function trigger device. When the soil moisture content is high, the spatial GCN module output contributes more to the predicted target. The feature enhancement fusion module output is given by
O f u s i o n = f O 1 + O 2 + O 1 tanh O 2 ,
where f ( · ) is an activation function, such as t a n h or s i g m o i d . Additionally, ⊙ represents element multiplication.
This gating mechanism enables context-aware weighting of spatial patterns, allowing the model to emphasize different features under dry versus wet watershed conditions.

2.3.3. Graph-LSTM Network Module

In this section, we present a dynamic time-based Graph-LSTM network for predicting streamflow [43]. At each time step, the spatial GCN network extracts the feature matrix. Based on the similarity of the features, we connect all the feature matrices between adjacent moments to form a sequence graph. The Graph-LSTM then takes the spatial features extracted from the rainfall-runoff data as input and generates a spatiotemporal hydrological feature matrix with time series features to predict streamflow.
Figure 5 shows how the hidden layer status of Graph-LSTM is updated in the time dimension. The update equation is given by
h i t = Graph - LSTM h i 1 t , u N i W h u t 1 ,
where W is a hyperparameter that can be learned, N i is the number of h i t units related to the state of the hidden layer at time t. Specifically, each Graph-LSTM unit can be decomposed as follows:
i t = σ W i x t + j K U i m t , j h j + b i ,
o t = σ W o x t + j K U o m t , j h j + b i ,
f t , j = σ W f x t + U f m t , j h j + b f ,
c ˜ t = tanh W c x t + j K U c m t , j h j + b c ,
c t = i t c ˜ t + j K f t , j c j ,
h t = o t tanh c t .
The structure of the Graph-LSTM network unit is similar to that of a conventional LSTM network, consisting of input gates, forget gates, and output gates. In Formulas (9) to (11), i t represents the input gate, o t represents the output gate, and f t , j represents the forget gate. The main difference is that a Graph-LSTM unit consists of multiple inputs from the previous time step, where i t and o t depend on the state of all Graph-LSTM units k time steps before the current moment, and m t , j indicates that the current moment t received information from the j-th node in the previous time step. The forget gate f t , j only depends on the corresponding gate in the previous state. c t ˜ and c t represent the intermediate calculation results in the Graph-LSTM unit. Finally, i t , f t , j , and o t are combined to generate the hidden layer state h t .

2.4. Output Prediction Module

Model output: the predicted streamflow value at the outlet of the watershed. Based on the law of flood formation and movement, the future flood development can be forecasted using past and real-time hydrometeorological data. In small- and medium-sized watersheds, the time from rainfall landing to the outlet of the watershed—known as the concentration time—is typically 6–8 h [44]. Therefore, the short-term flood prediction model output is set to the forecast period 1–8 h streamflow, which aligns with both operational forecasting needs and the watershed’s physical response characteristics.
The final output is obtained through a linear projection layer:
y ^ t + 1 : t + τ = W out H lstm + b out ,
where y ^ t + 1 : t + τ R τ represents the predicted streamflow for the next τ hours (with τ = 1 , , 8 ), H lstm is the final hidden state from the Graph-LSTM module, and W out , b out are learnable parameters.

3. Case Study

3.1. Data Source and Description

The geographical contexts and monitoring networks of the four study watersheds are presented in Figure 6. We selected four study watersheds for our research: Changhua watershed, Pingyao watershed, Chenhe watershed, and Daheba watershed. Of these, Changhua watershed and Pingyao watershed are humid watersheds, while the Chenhe watershed and Daheba watershed are semi-arid and semi-humid watersheds. Floods in Changhua and Pingyao watersheds are mostly caused by plum and typhoon rains, resulting in high peak heights. Meanwhile, Chenhe and Daheba watersheds, located at the southern foot of the Qinling Mountains, experience floods mainly due to local heavy rainfall.
In order to evaluate the performance of our proposed spatio-temporal graph neural network prediction model in streamflow forecasting, we conducted experiments using data from the Changhua watershed (1998–2010), the Chenhe watershed (2003–2010), the Pingyao watershed (2000–2012), and the Daheba watershed (2002–2014). A detailed description of the experimental data is presented in Table 2. The hydrological data for the four watersheds were collected from streamflow gauging stations (providing streamflow, evaporation, and soil moisture data) and rainfall stations (providing precipitation data). The granularity of each data collection was hourly.
The hydrological time series contained a small proportion of missing or abnormal values (e.g., negative flow readings, instrument outliers). Across the four watersheds, missing data accounted for less than 0.5% of the total hourly records. These gaps were filled using linear interpolation for periods shorter than 6 h; longer gaps were treated as missing and excluded from model input. Abnormal values (identified as readings beyond ±3 standard deviations from the moving average) were replaced by the median of the preceding 24 h. All processed series were visually inspected to ensure physical plausibility before being used for training and testing.
There were some incomplete instances or outlier values in the hydration data. Additionally, to analyze the effect of features on the output under different watersheds, we performed some statistical analysis. The statistical analysis of the flow data for these four datasets is shown in Table 3. The data sample size of the four datasets is different, with the Changhua dataset having a relatively small sample size, but more effective training samples. Moreover, the median streamflow features in the four datasets are smaller than their average, with the median flow feature of the Daheba dataset being much smaller than its average.
Furthermore, we carried out a statistical analysis of the rainfall features for these four datasets, the results of which are shown in Table 4. The maximum, average, and standard deviation of rainfall in the Changhua and Pingyao datasets are significantly larger than that of the Chenhe and Daheba datasets. Moreover, the median rainfall features in each dataset are less than their average.

3.2. Baseline Models

To verify the superiority and applicability of the ST-GNN in streamflow prediction, we conducted comparative experiments on four datasets using the following methods:

3.2.1. Benchmarks

We used five baseline models and the proposed models for the comparative experiments:
  • LSTM: long short-term memory, which is specifically designed to solve the long-term dependency problem of the general RNN [45] and suitable for prediction tasks of longer time series.
  • STA-LSTM: Spatio-temporal attention long short-term memory model [46]. This uses an attention mechanism and LSTM to predict floods.
  • GCN: convolutional neural networks. ZHAO et al. [47] use GCN for feature extraction and additional decode module for gated recurrent unit networks.
  • ST-GCN: spatial-temporal graph neural network, which is popular in recent time series data prediction tasks such as groundwater levels prediction [48].
  • AGCLSTM: Graph convolution-based spatial-temporal attention LSTM [49], which uses ST-GCN combined with attention mechanism to predict floods.
  • ST-GNN: spatial-temporal graph neural network model proposed by us.
All baseline models (LSTM, STA-LSTM, GCN, ST-GCN, AGCLSTM) were configured with only historical streamflow and rainfall data as inputs, following standard hydrological forecasting benchmarks. The auxiliary variables used exclusively by our ST-GNN—soil moisture, evaporation, and future rainfall trends—were not included in the baseline model inputs. This configuration was maintained to ensure fair comparison while demonstrating ST-GNN’s unique capability to integrate heterogeneous data sources through its specialized architectural components (feature enhancement fusion layer and Graph-LSTM module).

3.2.2. Model Variants

To explore the effects of the modules in the proposed ST-GNN model, we compare them to different variants:
  • ST-GNN-rglstm: The variant model of ST-GNN that removes Graph-LSTM network module.
  • ST-GNN-rgcn: The variant model of ST-GNN that removes GCN module.
  • ST-GNN-rfi: The variant model of ST-GNN that removes future rainfall input module.
  • ST-GNN-rfe: The variant model of ST-GNN that removes feature enhancement fusion module.
Note that all variant models (ST-GNN-rglstm, ST-GNN-rgcn, ST-GNN-rfi, ST-GNN-rfe) maintain the same multi-source input configuration as the full ST-GNN model, including soil moisture, evaporation, and future rainfall data where applicable. The ablation studies therefore isolate the architectural contributions rather than input differences.

3.3. Evaluation Indices

We used three evaluation metrics to measure the performance of the models.
  • Root mean square error (RMSE)
RMSE = 1 N i = 1 N ( y i o b s y i p r e ) 2
RMSE is measured as the average deviation between the predicted and real values in the regression tasks.
  • Deterministic coefficient (DC)
D C = 1 i = 1 N y i o b s y i p r e 2 i = 1 N y i o b s y ¯ p r e 2 .
The deterministic coefficient (DC), which is mathematically equivalent to the coefficient of determination ( R 2 ) and the Nash–Sutcliffe efficiency (NSE) in hydrological contexts, quantifies the model’s efficiency in explaining the variance of the observed data. We retain the notation DC for consistency with prior hydrological forecasting literature [46,49].
  • The relative difference of RMSE (%)
R M S E ( % ) = R M S E m o d e l 1 R M S E m o d e l 2 R M S E m o d e l 1 ,
where y i o b s is the measured value at time i, y i p r e is the predicted value at time i, y ¯ p r e is the average predicted value, y ¯ o b s is the average measured value, and N is the number of test samples.

3.4. Experimental Methods

In our experimental setup, each prediction model was trained on a single NVIDIA GeForce TITAN Xp GPU, using Python 3.6 as the programming language and PyTorch 1.8 as the deep learning library. We built a 4-layer GCN with 64 hidden units per layer and a Graph-LSTM module with 128 hidden units. The models were trained for a maximum of 300 epochs with early stopping applied if the validation loss did not improve for 30 consecutive epochs. The Adam optimizer was used with a learning rate of 0.001 and a batch size of 256. Under this configuration, a complete training run for one watershed (including validation) typically took 2–3 h. All experiments were repeated five times with different random seeds, and the reported results are the average over these independent runs.

4. Results and Discussion

4.1. Comprehensive Evaluation of ST-GNN Predictive Performance and Discussion

This subsection presents a rigorous multi-step, multi-watershed comparative evaluation of the proposed ST-GNN model against five state-of-the-art baseline models. The evaluation spans forecast lead times from 1 to 8 h across four hydrologically diverse watersheds. The primary objective is to quantify the performance gains attributable to ST-GNN’s integrated architecture—specifically its dynamic graph learning, feature enhancement fusion, and hybrid Graph-LSTM components. As detailed below, the results consistently demonstrate the superior predictive accuracy, stability, and adaptability of our framework.

4.1.1. Comprehensive Evaluation of ST-GNN Predictive Performance

To comprehensively evaluate the predictive performance of our proposed ST-GNN model, we present detailed multi-step prediction results based on RMSE and DC metrics across four distinct hydrological datasets, as shown in Table 5, Table 6, Table 7 and Table 8. The experimental results demonstrate that our ST-GNN model consistently outperforms all baseline models, including traditional methods (LSTM, GCN) and spatial-temporal variants (STA-LSTM, ST-GCN, AGCLSTM).
We first examine the results for the humid Changhua watershed (Table 5). In this high-rainfall regime, all models incorporating spatial information (STA-LSTM, GCN, ST-GCN, AGCLSTM, ST-GNN) outperform the sequence-only LSTM baseline, confirming the necessity of capturing watershed connectivity. Among these, ST-GNN achieves the lowest RMSE and highest DC values across all eight prediction steps.
The performance comparison reveals several key findings: first, spatial-temporal models (STA-LSTM, AGCLSTM, ST-GCN, ST-GNN) consistently surpass traditional approaches, highlighting the importance of capturing both spatial and temporal dependencies. Second, our ST-GNN demonstrates superior performance compared to the best-performing baseline model (AGCLSTM). Compared to AGCLSTM, ST-GNN achieves relative improvements in RMSE and DC of +0.79% and +3.75% in Changhua (humid), +11.48% and +4.46% in Chenhe (semi-arid/semi-humid), +8.25% and +3.26% in Pingyao (humid), and +1.48% and +6.04% in Daheba (semi-arid/semi-humid), respectively (calculated at the 6-h prediction horizon). Notably, the improvement is particularly significant in semi-arid/semi-humid watersheds (Chenhe and Daheba), suggesting our model’s enhanced capability in handling watersheds with low water content.
The statistical significance of these performance improvements was formally verified. As detailed in Supplementary Table S1, paired tests comparing ST-GNN against each baseline model for the 6 h ahead prediction—a critical mid-range forecasting horizon—consistently yielded p-values below 0.05 for both RMSE and DC across all four watersheds. This confirms that the superior performance of ST-GNN is not due to random variation but is statistically significant.

4.1.2. Discussion

The consolidated results from Table 5, Table 6, Table 7 and Table 8 unequivocally demonstrate the superior and statistically significant performance of the proposed ST-GNN framework across diverse hydrological regimes. Three key contributions emerge from this comparative analysis:
First, ST-GNN consistently outperforms not only sequence-based models (LSTM) but also advanced spatio-temporal baselines. Its average RMSE reduction of 6.5% over the strong ST-GCN baseline [48] highlights the efficacy of its integrated architecture. This improvement is particularly notable because ST-GCN itself represents a significant advance over pure temporal or spatial models, as shown in prior studies [47].
Second, the performance gain is most substantial in semi-arid/semi-humid watersheds (Chenhe and Daheba). This finding directly addresses a known challenge in hydrological forecasting: models often struggle under low soil moisture conditions where rainfall-runoff relationships are highly nonlinear and threshold-dependent [38,39]. The feature enhancement fusion module in ST-GNN, by explicitly incorporating soil moisture and evaporation, provides a mechanistic advantage here. It enables the model to dynamically modulate spatial feature importance based on watershed state, a capability lacking in attention-based models like that of Yang et al. [29] which focus primarily on feature weighting without explicit physical state integration.
Third, the model’s stability over extended lead times (1–8 h) surpasses that of recurrent network baselines like AGCLSTM [49]. The slower degradation of the DC metric (see Section 4.3) suggests that the Graph-LSTM module, with its dual capacity for spatial aggregation and temporal memory, mitigates error propagation more effectively than standard LSTMs or their spatially attentive variants. This aligns with the motivation behind recent hybrid architectures [33] but demonstrates improved generalization across watershed types.
In summary, these results validate the core hypotheses of this work: that a synergistic design combining (1) dynamically learned graphs (addressing the limitation of static graphs in [24]), (2) physics-informed feature fusion (extending beyond purely data-driven attention [29]), and (3) joint spatio-temporal modeling (enhancing upon sequential processing [49]) is crucial for advancing the accuracy and robustness of streamflow prediction, especially under heterogeneous climatic conditions.

4.2. Optimization Analysis of ST-GNN Input Timesteps

To systematically investigate the optimal input time step configuration for our ST-GNN model across watersheds with varying soil moisture conditions, we conducted extensive experiments evaluating 6 h prediction performance under different input time steps. As illustrated in Figure 7, we examined seven candidate input time steps ( k = [ 4 , 5 , 6 , 7 , 8 , 9 , 10 ] ) across all four watershed datasets. The RMSE results demonstrate a consistent concave pattern across all datasets, initially decreasing to a minimum point before subsequently increasing with longer input time steps. This characteristic pattern suggests the existence of an optimal trade-off between temporal context capture and noise accumulation.
Through rigorous analysis of these curves, we identified watershed-specific optimal configurations: k = 6 for Changhua (humid), k = 8 for Chenhe (semi-arid and semi-humid), k = 6 for Pingyao (humid), and k = 7 for Daheba (semi-arid and semi-humid). The complementary DC metric results presented in Figure 7b fully corroborate these findings, showing peak performance at identical input time steps. These optimal values reflect fundamental hydrological differences: semi-arid and semi-humid watersheds benefit from longer temporal context (k = 7–8), while humid watersheds achieve optimal performance with shorter input sequences ( k = 6 ). This empirical evidence provides valuable guidance for model configuration in different hydrological regimes.
This finding of watershed-specific optimal input lengths corroborates the concept of hydrological memory being variable across climates [7] and suggests that a one-size-fits-all input window, commonly used in many LSTM-based studies [8], may be suboptimal. Our adaptive framework allows this parameter to be tuned based on watershed characteristics.

4.3. Analysis of ST-GNN’s Long-Term Prediction Stability

The study specifically examines the ST-GNN model’s capability in long-term predictions by analyzing its performance variation across increasing prediction horizons (1–8 h) in four watershed datasets. As shown in Figure 8, the model demonstrates consistent stability, with comparable RMSE increase rates across all watershed types (Figure 8a). This stability is further confirmed by DC metric results (Figure 8b), indicating the model’s robustness regardless of initial watershed conditions—performing equally well in both humid (Changhua, Pingyao) and semi-arid/semi-humid (Chenhe, Daheba) watersheds.
To quantify error accumulation, we calculated the average decline rate of DC per hour between the 1-h and 8-h predictions: ST-GNN showed slower decline (Changhua: 0.0142/h, Chenhe: 0.0141/h, Pingyao: 0.0144/h, Daheba: 0.0115/h) than AGCLSTM (0.0150/h, 0.0153/h, 0.0157/h, 0.0130/h, respectively), confirming its superior stability over longer lead times.
The minimal performance degradation observed across extended prediction steps highlights ST-GNN’s superior temporal modeling capacity. Particularly noteworthy is the model’s consistent behavior in different hydrological regimes, suggesting its feature extraction mechanisms effectively compensate for the accumulated uncertainty in long-term predictions.
The quantified slower decline rate of ST-GNN’s DC metric compared to AGCLSTM (0.0142/h vs. 0.0150/h in Changhua) demonstrates its enhanced capacity for temporal dependency modeling. This addresses a common criticism of complex models having higher short-term accuracy but poor long-term stability [15].

4.4. Ablation Study on Key Modules of ST-GNN Model

To systematically investigate the impact of different spatio-temporal components on the predictive performance of our ST-GNN model, we conducted comprehensive ablation studies by removing four key modules: the Graph-LSTM module, the spatial GCN module, the future rainfall input module, and the feature enhancement fusion module.
As detailed in Section 3.2.2, we evaluated four variant models: ST-GNN-rglstm (without Graph-LSTM), ST-GNN-rgcn (without spatial GCN), ST-GNN-rfi (without future rainfall input), and ST-GNN-rfe (without feature enhancement fusion). The comparative analysis was performed using 6-step predictions across all four watershed datasets.
The experimental results (Figure 9a, Figure 10a, Figure 11a and Figure 12a) reveal three significant findings: First, the complete ST-GNN model consistently achieves the lowest RMSE values, demonstrating its superior predictive accuracy. Second, the removal of the future rainfall input module (ST-GNN-rfi) results in the most substantial performance degradation, particularly in semi-arid (Pingyao) and semi-humid (Changhua) watersheds, highlighting this module’s critical role in rainfall-runoff modeling. Third, the relative importance of other modules follows the order: feature enhancement fusion (most important) > Graph-LSTM > spatial GCN (least important), as evidenced by the progressively smaller RMSE increases when these modules are removed.
Complementary results for the DC metric (Figure 9b, Figure 10b, Figure 11b and Figure 12b) show identical trends, further validating these findings. The consistent performance patterns across both metrics and multiple watersheds underscore the robustness of our architectural design and the synergistic effects of the integrated modules.
The critical role of the future rainfall input module, especially in semi-arid regions, underscores the value of integrating even simple forecast products, a practice supported by operational hydrology literature [8]. The hierarchy of module importance (fusion > Graph-LSTM > GCN) provides clear guidance for model simplification or transfer learning in data-scarce settings.

4.5. Watershed Adaptability Analysis of Feature Enhancement Fusion Module

Figure 13 presents a comprehensive evaluation of the feature enhancement fusion module’s effectiveness across four distinct watershed types. The analysis of RMSE(%) improvements at 1–8 h prediction steps reveals striking contrasts: humid watersheds show more modest improvements (Changhua: 7.5%, Pingyao: 8.7%), while semi-arid and semi-humid watersheds demonstrate remarkable gains (Chenhe: 23.6%, Daheba: 21.2%). This substantial performance gap—nearly threefold between the most and least responsive watersheds—underscores the module’s exceptional value in water-variable environments.
The progressive enhancement pattern across prediction horizons suggests the module’s dynamic adaptation capability. In semi-arid and semi-humid watersheds, where soil moisture exhibits greater temporal variability, the fusion mechanism appears to develop increasingly effective feature representations over extended prediction periods. This temporal accumulation effect results in the particularly strong performance at longer prediction steps (6–8 h), where conventional models typically struggle most with moisture-related prediction challenges.
The dramatically larger improvement from the feature fusion module in semi-arid watersheds (23.6%) compared to humid ones (7.5%) provides empirical evidence for the module’s core function: it effectively encodes the time-varying sensitivity of runoff generation to rainfall, which is a fundamental concept in catchment hydrology [38,39]. This makes ST-GNN particularly valuable for flood forecasting in climate-sensitive regions.

5. Conclusions

This study presents a novel spatial-temporal graph neural network (ST-GNN) framework that addresses critical challenges in streamflow forecasting by integrating dynamic graph learning, physically informed feature enhancement, and hybrid Graph-LSTM architecture. The symmetry between the time and space modules is a key architectural feature of our proposed network. The proposed approach directly tackles the limitations of existing methods identified in the introduction, particularly the inability to capture dynamic spatiotemporal correlations and integrate domain-specific physical knowledge.
Our comprehensive experimental evaluation across four hydrologically diverse Chinese watersheds demonstrates the superior performance of the ST-GNN framework. Key findings include: (1) ST-GNN consistently outperforms state-of-the-art baseline models (LSTM, STA-LSTM, GCN, ST-GCN, AGCLSTM) across all prediction horizons (1–8 h), with particularly significant improvements in semi-arid regions; (2) The performance gains are statistically significant and robust, as confirmed by rigorous significance testing and multi-run stability analysis; (3) Ablation studies quantify the critical contribution of each architectural component, revealing a clear hierarchy of importance; (4) The model effectively adapts to varying watershed initial states, demonstrating its practical applicability under different hydrological conditions.
The practical value of our approach for local water resource managers and stakeholders is substantial. The model’s 1–8 h forecast capability, enhanced accuracy during flood peaks, and ability to integrate multi-source data (real-time soil moisture, evaporation, future rainfall trends) provide operational support for flood warning systems, reservoir operation decisions, and emergency response planning in medium-small watersheds.
We acknowledge two key limitations of this study: first, the watersheds examined are located within specific geographic and climatic regions of China. While providing valuable insights for similar regions, this may limit immediate generalizability to other hydroclimatic zones with fundamentally different rainfall-runoff mechanisms. Second, the dynamic graph learning mechanism is demonstrated and validated within individual watersheds but not across different hydroclimatic regions. The watersheds examined do not share common extreme events or synchronous seasonal transitions; therefore, the model’s ability to learn transferable, universal spatial correlation patterns has not been tested.
Future research will focus on enhancing model applicability through transfer learning and domain adaptation techniques, enabling the ST-GNN framework to be effectively applied to ungauged or data-scarce watersheds—a significant practical challenge in many regions worldwide. In addition, applying transfer learning and domain adaptation techniques to evaluate and enhance the cross-watershed generalizability of the learned dynamic spatial relationships.
In conclusion, this work demonstrates the significant value of integrating physical hydrological principles with advanced graph neural networks. The ST-GNN framework advances the state of the art in streamflow forecasting while providing a robust foundation for practical applications. Continued research to enhance model transferability across diverse hydrological settings will further increase its impact on water resources management and disaster mitigation efforts.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/sym18020240/s1.

Author Contributions

Conceptualization, L.Y. and J.F.; methodology, L.Y. and J.F.; software, L.Y.; validation, L.Y., T.H., and D.S.; formal analysis, L.Y.; investigation, H.Z.; resources, X.Z.; data curation, J.L.; writing-original draft preparation, L.Y. and J.F.; writing-review and editing, L.Z.; visualization, Y.L.; supervision, L.Y. and J.F. All authors have read and agreed to the published version of the manuscript.

Funding

This research has been supported by the National Key Research and Development Program of China (2023YFC3209203), and the National Natural Science Foundation of China (Grant No. 62306007).

Data Availability Statement

The original contributions presented in the study are included in the article/Supplementary Material, further inquiries can be directed to the corresponding author.

Conflicts of Interest

Author Dacheng Shan was employed by the company Nari Group Corporation (State Grid Electric Power Research Institute) and Nanjing NARI Information and Communication Technology Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Zhou, G.; Zhao, Q.; Zhang, S.; Zhang, D.; Li, C. Rolling forecast of snowmelt floods in data-scarce mountainous regions using weather forecast products to drive distributed energy balance hydrological model. J. Hydrol. 2024, 637, 131384. [Google Scholar] [CrossRef]
  2. Izadi, A.; Zarei, N.; Nikoo, M.R.; Al-Wardy, M.; Yazdandoost, F. Exploring the potential of deep learning for streamflow forecasting: A comparative study with hydrological models for seasonal and perennial rivers. Expert Syst. Appl. 2024, 252, 124139. [Google Scholar] [CrossRef]
  3. Kim, S.; Matsumi, Y.; Pan, S.; Mase, H. A real-time forecast model using artificial neural network for after-runner storm surges on the Tottori Coast, Japan. Ocean Eng. 2016, 122, 44–53. [Google Scholar] [CrossRef]
  4. Tareke, K.A.; Awoke, A.G. Hydrological drought forecasting and monitoring system development using artificial neural network (ANN) in Ethiopia. Heliyon 2023, 9, e13287. [Google Scholar] [CrossRef]
  5. Araghinejad, S.; Azmi, M.; Kholghi, M. Application of artificial neural network ensembles in probabilistic hydrological forecasting. J. Hydrol. 2011, 407, 94–104. [Google Scholar]
  6. Zhang, J.; Zhu, Y.; Zhang, X.; Ye, M.; Yang, J. Developing a Long Short-Term Memory (LSTM) based model for predicting water table depth in agricultural areas. J. Hydrol. 2018, 561, 918–929. [Google Scholar] [CrossRef]
  7. Gao, S.; Huang, Y.; Zhang, S.; Han, J.; Wang, G.; Zhang, M.; Lin, Q. Short-term runoff prediction with GRU and LSTM networks without requiring time step optimization during sample generation. J. Hydrol. 2020, 589, 125188. [Google Scholar] [CrossRef]
  8. Ni, L.; Wang, D.; Singh, V.P.; Wu, J.; Wang, Y.; Tao, Y.; Zhang, J. Streamflow and rainfall forecasting by two long short-term memory-based models. J. Hydrol. 2020, 583, 124296. [Google Scholar] [CrossRef]
  9. Deng, H.; Chen, W.; Huang, G. Deep insight into daily runoff forecasting based on a CNN-LSTM model. Nat. Hazards 2022, 113, 1675–1696. [Google Scholar] [CrossRef]
  10. Khorram, S.; Jehbez, N. A hybrid CNN-LSTM approach for monthly reservoir inflow forecasting. Water Resour. Manag. 2023, 37, 4097–4121. [Google Scholar] [CrossRef]
  11. Dehghani, A.; Moazam, H.M.Z.H.; Mortazavizadeh, F.; Ranjbar, V.; Mirzaei, M.; Mortezavi, S.; Ng, J.L.; Dehghani, A. Comparative evaluation of LSTM, CNN, and ConvLSTM for hourly short-term streamflow forecasting using deep learning approaches. Ecol. Inform. 2023, 75, 102119. [Google Scholar] [CrossRef]
  12. Malik, H.; Feng, J.; Shao, P.; Abduljabbar, Z.A. Improving flood forecasting using time-distributed CNN-LSTM model: A time-distributed spatiotemporal method. Earth Sci. Inform. 2024, 17, 3455–3474. [Google Scholar] [CrossRef]
  13. Jin, A.; Wang, Q.; Zhan, H.; Zhou, R. Comparative performance assessment of physical-based and data-driven machine-learning models for simulating streamflow: A case study in three catchments across the US. J. Hydrol. Eng. 2024, 29, 05024004. [Google Scholar] [CrossRef]
  14. Hu, X.; Shi, L.; Lin, G.; Lin, L. Comparison of physical-based, data-driven and hybrid modeling approaches for evapotranspiration estimation. J. Hydrol. 2021, 601, 126592. [Google Scholar] [CrossRef]
  15. Sabzipour, B.; Arsenault, R.; Troin, M.; Martel, J.L.; Brissette, F.; Brunet, F.; Mai, J. Comparing a long short-term memory (LSTM) neural network with a physically-based hydrological model for streamflow forecasting over a Canadian catchment. J. Hydrol. 2023, 627, 130380. [Google Scholar] [CrossRef]
  16. Scarselli, F.; Gori, M.; Tsoi, A.C.; Hagenbuchner, M.; Monfardini, G. The graph neural network model. IEEE Trans. Neural Netw. 2008, 20, 61–80. [Google Scholar] [CrossRef]
  17. Feng, Z.; Wang, R.; Wang, T.; Song, M.; Wu, S.; He, S. A comprehensive survey of dynamic graph neural networks: Models, frameworks, benchmarks, experiments and challenges. IEEE Trans. Knowl. Data Eng. 2025, 38, 26–46. [Google Scholar] [CrossRef]
  18. Liu, H.; Yang, D.; Liu, X.; Chen, X.; Liang, Z.; Wang, H.; Cui, Y.; Gu, J. Todynet: Temporal dynamic graph neural network for multivariate time series classification. Inf. Sci. 2024, 677, 120914. [Google Scholar] [CrossRef]
  19. Lu, G.; Bu, S. Online Power System Dynamic Security Assessment: A GNN–FNO Approach Learning From Multisource Spatial–Temporal Data. IEEE Trans. Ind. Inform. 2025, 21, 7598–7608. [Google Scholar] [CrossRef]
  20. Gu, W.; Sun, M.; Liu, B.; Xu, K.; Sui, M. Adaptive Spatio-Temporal Aggregation for Temporal Dynamic Graph-Based Fraud Risk Detection. J. Comput. Technol. Softw. 2024, 3, 1–5. [Google Scholar]
  21. Liu, J.; Liu, J.; Zhao, K.; Tang, Y.; Chen, W. TP-GNN: Continuous dynamic graph neural network for graph classification. In Proceedings of the 2024 IEEE 40th International Conference on Data Engineering (ICDE), Utrecht, The Netherlands, 13–16 May 2024; IEEE: New York, NY, USA, 2024; pp. 2848–2861. [Google Scholar]
  22. Wang, T.; Ding, Z.; Chang, Z.; Yang, X.; Chen, Y.; Li, M.; Xu, S.; Wang, Y. A novel graph neural network framework for resting-state functional MRI spatiotemporal dynamics analysis. Phys. A Stat. Mech. Its Appl. 2025, 669, 130582. [Google Scholar] [CrossRef]
  23. Molaei, S.; Niknam, G.; Ghosheh, G.O.; Chauhan, V.K.; Zare, H.; Zhu, T.; Pan, S.; Clifton, D.A. Temporal dynamics unleashed: Elevating variational graph attention. Knowl.-Based Syst. 2024, 299, 112110. [Google Scholar] [CrossRef]
  24. Liu, Y.; Hou, G.; Huang, F.; Qin, H.; Wang, B.; Yi, L. Directed graph deep neural network for multi-step daily streamflow forecasting. J. Hydrol. 2022, 607, 127515. [Google Scholar] [CrossRef]
  25. Kazadi, A.N.; Doss-Gollin, J.; Sebastian, A.; Silva, A. Flood prediction with graph neural networks. Clim. Change AI 2022. Available online: https://s3.us-east-1.amazonaws.com/climate-change-ai/papers/neurips2022/75/paper.pdf (accessed on 25 January 2026).
  26. Bai, T.; Tahmasebi, P. Graph neural network for groundwater level forecasting. J. Hydrol. 2023, 616, 128792. [Google Scholar] [CrossRef]
  27. Taghizadeh, M.; Zandsalimi, Z.; Nabian, M.A.; Shafiee-Jood, M.; Alemazkoor, N. Interpretable physics-informed graph neural networks for flood forecasting. Comput.-Aided Civ. Infrastruct. Eng. 2025, 40, 2629–2649. [Google Scholar] [CrossRef]
  28. Ashraf, I.; Strotherm, J.; Hermes, L.; Hammer, B. Physics-informed graph neural networks for water distribution systems. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 20–27 February 2024; Volume 38, pp. 21905–21913. [Google Scholar]
  29. Yang, B.; Chen, L.; Yi, B.; Li, S.; Leng, Z. Local Weather and Global Climate Data-Driven Long-Term Runoff Forecasting Based on Local–Global–Temporal Attention Mechanisms and Graph Attention Networks. Remote Sens. 2024, 16, 3659. [Google Scholar] [CrossRef]
  30. Li, J.; Hsu, K.l.; Jiang, A.L.; Zhang, Y.; Ye, A.; Sorooshian, S. Improving Rainfall-runoff Modeling Using an Attention-based Model. Authorea Prepr. 2025. [Google Scholar] [CrossRef]
  31. Yuan, W.; Yan, H. Enhancing runoff prediction with causal lag-aware attention and multi-scale fusion in transformer models. J. Hydrol. 2025, 664, 134369. [Google Scholar] [CrossRef]
  32. Sheng, Z.; Cao, Y.; Yang, Y.; Feng, Z.K.; Shi, K.; Huang, T.; Wen, S. Residual temporal convolutional network with dual attention mechanism for multilead-time interpretable runoff forecasting. IEEE Trans. Neural Netw. Learn. Syst. 2024, 36, 8757–8771. [Google Scholar] [CrossRef]
  33. Zhou, R. Multi-scale dynamic spatiotemporal graph attention network for forecasting karst spring discharge. J. Hydrol. 2025, 659, 133289. [Google Scholar] [CrossRef]
  34. Taccari, M.L.; Wang, H.; Nuttall, J.; Chen, X.; Jimack, P.K. Spatial-temporal graph neural networks for groundwater data. Sci. Rep. 2024, 14, 24564. [Google Scholar] [CrossRef]
  35. Xue, L.; Zhu, Y. Dynamic hydrological flow prediction with self-iterative spatiotemporal graph neural network: Modeling long-and short-period topological dynamics. J. Hydrol. 2025, 663, 134122. [Google Scholar] [CrossRef]
  36. Jin, G.; Liang, Y.; Fang, Y.; Shao, Z.; Huang, J.; Zhang, J.; Zheng, Y. Spatio-temporal graph neural networks for predictive learning in urban computing: A survey. IEEE Trans. Knowl. Data Eng. 2023, 36, 5388–5408. [Google Scholar] [CrossRef]
  37. Benesty, J.; Chen, J.; Huang, Y.; Cohen, I. Pearson correlation coefficient. In Noise Reduction in Speech Processing; Springer: Berlin/Heidelberg, Germany, 2009; pp. 1–4. [Google Scholar]
  38. Mahanama, S.; Livneh, B.; Koster, R.; Lettenmaier, D.; Reichle, R. Soil moisture, snow, and seasonal streamflow forecasts in the United States. J. Hydrometeorol. 2012, 13, 189–203. [Google Scholar] [CrossRef]
  39. Koster, R.D.; Mahanama, S.P.; Livneh, B.; Lettenmaier, D.P.; Reichle, R.H. Skill in streamflow forecasts derived from large-scale estimates of soil moisture and snow. Nat. Geosci. 2010, 3, 613–616. [Google Scholar] [CrossRef]
  40. Atwood, J.; Towsley, D. Diffusion-convolutional neural networks. In Proceedings of the Advances in Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; pp. 1993–2001. [Google Scholar]
  41. Kipf, T. Semi-supervised classification with graph convolutional networks. arXiv 2016, arXiv:1609.02907. [Google Scholar]
  42. Li, Y.; Yu, R.; Shahabi, C.; Liu, Y. Diffusion convolutional recurrent neural network: Data-driven traffic forecasting. In Proceedings of the International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  43. Shu, X.; Zhang, L.; Sun, Y.; Tang, J. Host–parasite: Graph LSTM-in-LSTM for group activity recognition. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 663–674. [Google Scholar] [CrossRef]
  44. Xu, W.; Chen, J.; Zhang, X.J. Scale effects of the monthly streamflow prediction using a state-of-the-art deep learning model. Water Resour. Manag. 2022, 36, 3609–3625. [Google Scholar] [CrossRef]
  45. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  46. Ding, Y.; Zhu, Y.; Feng, J.; Zhang, P.; Cheng, Z. Interpretable spatio-temporal attention LSTM model for flood forecasting. Neurocomputing 2020, 403, 348–359. [Google Scholar] [CrossRef]
  47. Zhao, Q.; Zhu, Y.; Shu, K.; Wan, D.; Yu, Y.; Zhou, X.; Liu, H. Joint Spatial and Temporal Modeling for Hydrological Prediction. IEEE Access 2020, 8, 78492–78503. [Google Scholar] [CrossRef]
  48. Chen, L.; Zhang, D.; Xu, J.; Zhou, Z.; Jin, J.; Luan, J.; Wulamu, A. Enhancing the accuracy of groundwater level prediction at different scales using spatio-temporal graph convolutional model. Earth Sci. Inform. 2025, 18, 250. [Google Scholar]
  49. Feng, J.; Sha, H.; Ding, Y.; Yan, L.; Yu, Z. Graph convolution based spatial-temporal attention LSTM model for flood forecasting. In Proceedings of the 2022 International Joint Conference on Neural Networks (IJCNN), Padua, Italy, 18–23 July 2022; IEEE: New York, NY, USA, 2022; pp. 1–8. [Google Scholar]
Figure 1. Spatio-temporal graph neural network prediction model framework.
Figure 1. Spatio-temporal graph neural network prediction model framework.
Symmetry 18 00240 g001
Figure 2. Two methods of establishing rainfall-runoff graph.
Figure 2. Two methods of establishing rainfall-runoff graph.
Symmetry 18 00240 g002
Figure 3. Flood initiation states under different soil moisture conditions. (a) Under low soil moisture: rainfall (blue bars) produces minimal streamflow response (Black line) due to high infiltration. (b) Under high soil moisture: the same rainfall input triggers a pronounced flood peak because the soil is near saturation, resulting in rapid runoff generation. The contrast illustrates why soil moisture is a critical state variable for rainfall-runoff modeling.
Figure 3. Flood initiation states under different soil moisture conditions. (a) Under low soil moisture: rainfall (blue bars) produces minimal streamflow response (Black line) due to high infiltration. (b) Under high soil moisture: the same rainfall input triggers a pronounced flood peak because the soil is near saturation, resulting in rapid runoff generation. The contrast illustrates why soil moisture is a critical state variable for rainfall-runoff modeling.
Symmetry 18 00240 g003
Figure 4. Schematic of the feature enhancement fusion module. The module takes two inputs: O 1 (spatial features from the GCN module) and O 2 (processed watershed state features). A tanh-gating mechanism ( O 1 tanh ( O 2 ) ) dynamically weights the contribution of spatial features based on soil moisture content. The gated result is added to the original inputs, and an activation function f ( · ) produces the fused output O fusion . This design enables context-aware fusion of spatial patterns with watershed wetness conditions.
Figure 4. Schematic of the feature enhancement fusion module. The module takes two inputs: O 1 (spatial features from the GCN module) and O 2 (processed watershed state features). A tanh-gating mechanism ( O 1 tanh ( O 2 ) ) dynamically weights the contribution of spatial features based on soil moisture content. The gated result is added to the original inputs, and an activation function f ( · ) produces the fused output O fusion . This design enables context-aware fusion of spatial patterns with watershed wetness conditions.
Symmetry 18 00240 g004
Figure 5. Temporal updating mechanism of Graph-LSTM hidden states. At each time step t, the hidden state h i ( t ) of node i is updated by aggregating information from its neighboring nodes u N ( i ) (spatial aggregation) and from the previous hidden state h i ( t 1 ) (temporal memory). The Equations (9)–(14) detail the input, forget, and output gates that regulate information flow. The diagram highlights how Graph-LSTM jointly captures spatial dependencies (via neighborhood aggregation) and temporal dynamics (via recurrent connections).
Figure 5. Temporal updating mechanism of Graph-LSTM hidden states. At each time step t, the hidden state h i ( t ) of node i is updated by aggregating information from its neighboring nodes u N ( i ) (spatial aggregation) and from the previous hidden state h i ( t 1 ) (temporal memory). The Equations (9)–(14) detail the input, forget, and output gates that regulate information flow. The diagram highlights how Graph-LSTM jointly captures spatial dependencies (via neighborhood aggregation) and temporal dynamics (via recurrent connections).
Symmetry 18 00240 g005
Figure 6. Spatial distribution of the four study watersheds and their monitoring networks. (a) Changhua watershed: A humid watershed with six rainfall stations (blue circles) and 1 streamflow gauging station (red triangle). (b) Chenhe watershed: A semi-arid/semi-humid watershed with eight rainfall stations and one streamflow station. (c) Daheba watershed: A semi-arid watershed with 15 rainfall stations and 1 streamflow station. (d) Pingyao watershed: A humid watershed with seven rainfall stations and one streamflow station. Each map shows the watershed boundary (black line), river network (blue lines), and the spatial configuration of monitoring stations.
Figure 6. Spatial distribution of the four study watersheds and their monitoring networks. (a) Changhua watershed: A humid watershed with six rainfall stations (blue circles) and 1 streamflow gauging station (red triangle). (b) Chenhe watershed: A semi-arid/semi-humid watershed with eight rainfall stations and one streamflow station. (c) Daheba watershed: A semi-arid watershed with 15 rainfall stations and 1 streamflow station. (d) Pingyao watershed: A humid watershed with seven rainfall stations and one streamflow station. Each map shows the watershed boundary (black line), river network (blue lines), and the spatial configuration of monitoring stations.
Symmetry 18 00240 g006
Figure 7. Optimal input timestep configuration for ST-GNN: k = 6 for humid Watersheds vs. k = 7–8 for Semi-Arid and Semi-Humid Watersheds.
Figure 7. Optimal input timestep configuration for ST-GNN: k = 6 for humid Watersheds vs. k = 7–8 for Semi-Arid and Semi-Humid Watersheds.
Symmetry 18 00240 g007
Figure 8. Performance trends of ST-GNN across four watersheds over 8 prediction steps.
Figure 8. Performance trends of ST-GNN across four watersheds over 8 prediction steps.
Symmetry 18 00240 g008
Figure 9. Performance comparison of ST-GNN variant models in Changhua watershed: RMSE and DC evaluation across 6 prediction timesteps. Module ablation reveals future rainfall input as most critical component, followed by feature enhancement fusion.
Figure 9. Performance comparison of ST-GNN variant models in Changhua watershed: RMSE and DC evaluation across 6 prediction timesteps. Module ablation reveals future rainfall input as most critical component, followed by feature enhancement fusion.
Symmetry 18 00240 g009
Figure 10. Performance comparison of ST-GNN variant models in Chenhe watershed: RMSE and DC evaluation across 6 prediction timesteps. Module ablation reveals feature enhancement fusion as most critical component, followed by future rainfall input.
Figure 10. Performance comparison of ST-GNN variant models in Chenhe watershed: RMSE and DC evaluation across 6 prediction timesteps. Module ablation reveals feature enhancement fusion as most critical component, followed by future rainfall input.
Symmetry 18 00240 g010
Figure 11. Performance comparison of ST-GNN variant models in Pingyao watershed: RMSE and DC evaluation across 6 prediction timesteps. Module ablation reveals future rainfall input as most critical component, followed by feature enhancement fusion.
Figure 11. Performance comparison of ST-GNN variant models in Pingyao watershed: RMSE and DC evaluation across 6 prediction timesteps. Module ablation reveals future rainfall input as most critical component, followed by feature enhancement fusion.
Symmetry 18 00240 g011
Figure 12. Performance comparison of ST-GNN variant models in Daheba watershed: RMSE and DC evaluation across 6 prediction timesteps. Module ablation reveals feature enhancement fusion as most critical component, followed by future rainfall input.
Figure 12. Performance comparison of ST-GNN variant models in Daheba watershed: RMSE and DC evaluation across 6 prediction timesteps. Module ablation reveals feature enhancement fusion as most critical component, followed by future rainfall input.
Symmetry 18 00240 g012
Figure 13. Analysis of prediction performance improvement by feature enhancement fusion module across watersheds. This figure compares the RMSE(%) improvement of the feature enhancement fusion module between humid watersheds (Changhua, Pingyao) and semi-arid/semi-humid watersheds (Chenhe, Daheba), where the latter show significantly greater enhancement (Chenhe 23.6%, Daheba 21.2%) compared to the former (Changhua 7.5%, Pingyao 8.7%).
Figure 13. Analysis of prediction performance improvement by feature enhancement fusion module across watersheds. This figure compares the RMSE(%) improvement of the feature enhancement fusion module between humid watersheds (Changhua, Pingyao) and semi-arid/semi-humid watersheds (Chenhe, Daheba), where the latter show significantly greater enhancement (Chenhe 23.6%, Daheba 21.2%) compared to the former (Changhua 7.5%, Pingyao 8.7%).
Symmetry 18 00240 g013
Table 1. Summary of the core modules in the proposed ST-GNN architecture. N: number of stations, F: number of input features, D: GCN hidden dimension, D h : LSTM hidden dimension, T: input sequence length, τ : forecast lead time.
Table 1. Summary of the core modules in the proposed ST-GNN architecture. N: number of stations, F: number of input features, D: GCN hidden dimension, D h : LSTM hidden dimension, T: input sequence length, τ : forecast lead time.
ModuleFunction & DescriptionInput/Output Specifications
Graph ConstructorConstructs adjacency matrices based on: (1) Physical distance ( A D , Equation (1)), (2) Statistical correlation ( A T , Equation (2)). Final adjacency: A = Normalize ( A D + λ A T ) Input: Station coordinates;
Historical features ( T × N × F )
Output:  A R N × N
Spatial GCNPerforms K-layer graph convolution (Equation (6)) using diffusion convolution. Aggregates spatial information from neighboring stations with summation across diffusion steps.Input:  X R N × F , A R N × N
Output:  H ( K ) R N × D
Parameters: K layers, hidden dim D
Feature FusionEnhances spatial GCN features O 1 (where O 1 = H ( K ) ) with watershed state features O 2 via gating mechanism: O fusion = f ( O 1 + O 2 + O 1 tanh ( O 2 ) ) (Equation (7)). Here, O 2 is obtained by applying a fully connected network (FCN) to the concatenated watershed state features X s t a t e from Equation (3).Input:  O 1 = H ( K ) R N × D ,
O 2 = FCN ( X s t a t e ) R N × D
Output:  O fusion R N × D
Graph-LSTMProcesses temporal sequence of fused features { O fusion } along with future rainfall information to capture spatiotemporal dependencies. Each LSTM unit (Equations (8)–(14)) aggregates information from neighboring nodes and incorporates future rainfall trends X f u t u r e (from Equation (4)) through an additional input gate.Input:  { O fusion } R T × N × D ,
FCN ( X f u t u r e ) R N × D h
Output:  H lstm R N × D h
Parameters: Hidden dim D h
Output LayerLinear projection for multi-step streamflow forecasts: y ^ t + 1 : t + τ = W out H lstm + b out Input:  H lstm R N × D h
Output:  y ^ R τ
Parameters:  W out R τ × D h , b out R τ
Table 2. Detailed description of the data for the four watersheds.
Table 2. Detailed description of the data for the four watersheds.
WatershedType of StationsCountTime GranularityTraining SetTest Set
ChanghuaHydrological station11 sample/h74831871
Rainfall station6
ChenheHydrological station11 sample/h14,7963699
Rainfall station8
PingyaoHydrological station11 sample/h12,3243081
Rainfall station7
DahebaHydrological station11 sample/h10,2472561
Rainfall station15
Table 3. Statistical indicators of the streamflow of the four watershed datasets.
Table 3. Statistical indicators of the streamflow of the four watershed datasets.
Data SetNumber of SamplesMaxMinAverageMedianSD
Changhua data set937121000.58146.6580.32202.49
Chenhe data set18,49617400.6936.9213.5078.77
Pingyao data set15,4066304.2297.8562.93106.22
Dahaba data set12,80933001.4085.7325.96209.07
Table 4. Statistical indicators of the rainfall of the four watershed datasets.
Table 4. Statistical indicators of the rainfall of the four watershed datasets.
Data SetNumber of SamplesMaxMinAverageMedianSD
Changhua data set937136.910.000.730.002.06
Chenhe data set18,4967.420.000.200.000.57
Pingyao data set15,40632.000.000.700.151.61
Dahaba data set12,80914.480.000.250.000.86
Table 5. Comparison of RMSE and DC metrics across 1–8 prediction steps for different models in Changhua watershed dataset, with ST-GNN’s superior performance highlighted in bold. Values represent the mean over 5 independent runs with different random seeds.
Table 5. Comparison of RMSE and DC metrics across 1–8 prediction steps for different models in Changhua watershed dataset, with ST-GNN’s superior performance highlighted in bold. Values represent the mean over 5 independent runs with different random seeds.
ModelsPredict Step1 h2 h3 h4 h5 h6 h7 h8 h
LSTMRMSE78.0279.8281.8583.9985.6188.6792.0896.98
DC0.9610.9520.9450.9330.9160.8930.8710.842
STA-LSTMRMSE74.3275.5476.7177.7280.6183.3386.0392.41
DC0.9750.9680.9580.9480.9280.9180.9010.867
GCNRMSE78.3378.9280.7582.7984.6187.3791.0895.48
DC0.9660.9520.9440.9320.9120.9030.8890.842
ST-GCNRMSE75.3276.5477.7178.7282.6185.3388.0395.41
DC0.9710.9630.9510.9440.9210.9140.8920.853
AGCLSTMRMSE73.8274.5476.2177.2279.6183.3385.0390.41
DC0.9780.9700.9610.9500.9320.9200.9060.874
ST-GNNRMSE72.3573.8275.6577.9580.6682.6785.9891.01
DC0.9810.9720.9650.9530.9360.9230.9110.882
Table 6. Comparison of RMSE and DC metrics across 1–8 prediction steps for different models in Chenhe watershed dataset, with ST-GNN’s superior performance highlighted in bold. Values represent the mean over 5 independent runs with different random seeds.
Table 6. Comparison of RMSE and DC metrics across 1–8 prediction steps for different models in Chenhe watershed dataset, with ST-GNN’s superior performance highlighted in bold. Values represent the mean over 5 independent runs with different random seeds.
ModelsPredict Step1 h2 h3 h4 h5 h6 h7 h8 h
LSTMRMSE60.3261.8263.3265.9167.6268.6172.0276.98
DC0.9460.9320.9270.9210.9050.8730.8510.822
STA-LSTMRMSE65.3266.5467.7168.7270.6172.3375.0378.41
DC0.9350.9280.9200.9120.9010.8850.8650.838
GCNRMSE60.6261.3263.1464.9967.6468.6771.9175.88
DC0.9450.9350.9250.9220.8990.8720.8490.821
ST-GCNRMSE58.6559.3561.3263.2365.6166.6470.1274.51
DC0.9420.9330.9140.9050.8920.8710.8570.831
AGCLSTMRMSE62.3263.5464.7165.7267.6169.3372.0376.41
DC0.9380.9300.9220.9150.9020.8880.8690.840
ST-GNNRMSE55.3555.8257.6358.7559.8961.3764.9468.01
DC0.9550.9420.9310.9180.9060.8930.8710.842
Table 7. Comparison of RMSE and DC metrics across 1–8 prediction steps for different models in Pingyao watershed dataset, with ST-GNN’s superior performance highlighted in bold. Values represent the mean over 5 independent runs with different random seeds.
Table 7. Comparison of RMSE and DC metrics across 1–8 prediction steps for different models in Pingyao watershed dataset, with ST-GNN’s superior performance highlighted in bold. Values represent the mean over 5 independent runs with different random seeds.
ModelsPredict Step1 h2 h3 h4 h5 h6 h7 h8 h
LSTMRMSE47.0548.2249.3951.4152.7453.5256.1860.04
DC0.9510.9420.9350.9230.9060.8840.8620.833
STA-LSTMRMSE44.8245.6846.9148.2249.4551.1753.2156.26
DC0.9600.9520.9430.9310.9150.9020.8770.848
GCNRMSE47.3748.4249.5451.2152.4153.5256.2659.84
DC0.9500.9410.9340.9260.9080.8880.8610.831
ST-GCNRMSE45.3146.4247.5550.2251.4553.1755.2158.26
DC0.9520.9450.9380.9260.9060.8940.8650.836
AGCLSTMRMSE43.3244.1845.4147.2248.4550.1752.2154.26
DC0.9650.9570.9480.9350.9200.9080.8850.855
ST-GNNRMSE41.5141.8743.2244.0644.9246.0348.7151.01
DC0.9740.9610.9500.9360.9240.9110.8880.859
Table 8. Comparison of RMSE and DC metrics across 1–8 prediction steps for different models in Daheba watershed dataset, with ST-GNN’s superior performance highlighted in bold. Values represent the mean over 5 independent runs with different random seeds.
Table 8. Comparison of RMSE and DC metrics across 1–8 prediction steps for different models in Daheba watershed dataset, with ST-GNN’s superior performance highlighted in bold. Values represent the mean over 5 independent runs with different random seeds.
ModelsPredict Step1 h2 h3 h4 h5 h6 h7 h8 h
LSTMRMSE100.65102.97105.59108.35110.44114.38118.78125.10
DC0.8960.8870.8810.8700.8540.8320.8120.785
STA-LSTMRMSE97.4599.12101.85104.62107.25110.38114.72121.15
DC0.9050.8980.8900.8780.8630.8460.8300.808
GCNRMSE101.91103.51106.52109.86110.76113.31117.71124.44
DC0.8950.8810.8790.8620.8460.8310.8110.789
ST-GCNRMSE98.92100.52103.53106.71108.72112.35116.72123.11
DC0.8990.8920.8830.8650.8550.8360.8160.797
AGCLSTMRMSE96.1597.82100.55103.32105.95109.08113.42119.85
DC0.9100.9030.8950.8830.8680.8510.8350.813
ST-GNNRMSE94.0695.9798.35101.34104.86107.47111.77118.31
DC0.9140.9060.8990.8880.8720.8600.8490.822
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yan, L.; Shan, D.; Zhu, X.; Zheng, L.; Zhang, H.; Li, Y.; Li, J.; Hang, T.; Feng, J. Streamflow Prediction of Spatio-Temporal Graph Neural Network with Feature Enhancement Fusion. Symmetry 2026, 18, 240. https://doi.org/10.3390/sym18020240

AMA Style

Yan L, Shan D, Zhu X, Zheng L, Zhang H, Li Y, Li J, Hang T, Feng J. Streamflow Prediction of Spatio-Temporal Graph Neural Network with Feature Enhancement Fusion. Symmetry. 2026; 18(2):240. https://doi.org/10.3390/sym18020240

Chicago/Turabian Style

Yan, Le, Dacheng Shan, Xiaorui Zhu, Lingling Zheng, Hongtao Zhang, Ying Li, Jing Li, Tingting Hang, and Jun Feng. 2026. "Streamflow Prediction of Spatio-Temporal Graph Neural Network with Feature Enhancement Fusion" Symmetry 18, no. 2: 240. https://doi.org/10.3390/sym18020240

APA Style

Yan, L., Shan, D., Zhu, X., Zheng, L., Zhang, H., Li, Y., Li, J., Hang, T., & Feng, J. (2026). Streamflow Prediction of Spatio-Temporal Graph Neural Network with Feature Enhancement Fusion. Symmetry, 18(2), 240. https://doi.org/10.3390/sym18020240

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop