Next Article in Journal
Numerical Investigation on the Hydrodynamic Coefficients of Subsea Suspended Pipelines Under Unidirectional Currents
Previous Article in Journal
A Comparative Study of a Two-Dimensional Slope Hydrodynamic Model (TDSHM), Long Short-Term Memory (LSTM), and Convolutional Neural Network (CNN) Models for Runoff Prediction
Previous Article in Special Issue
Evaluation of Micrometeorological Models for Estimating Crop Evapotranspiration Using a Smart Field Weighing Lysimeter
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Precipitation Spatio-Temporal Forecasting in China via DC-CNN-BiLSTM

1
State Key Laboratory of Public Big Data, College of Computer Science and Technology, Guizhou University, Guiyang 550025, China
2
Guizhou East Century Science Technology Co., Ltd., Guiyang 550089, China
3
College of Information Science and Engineering, University of Jinan, Jinan 421010, China
4
College of Big Data and Information Engineering, Guizhou University, Guiyang 550025, China
*
Author to whom correspondence should be addressed.
Water 2025, 17(9), 1381; https://doi.org/10.3390/w17091381
Submission received: 12 March 2025 / Revised: 20 April 2025 / Accepted: 30 April 2025 / Published: 4 May 2025
(This article belongs to the Special Issue Advances in Crop Evapotranspiration and Soil Water Content)

Abstract

:
Accurate and reliable precipitation prediction remains a significant challenge due to an incomplete understanding of regional meteorological dynamics and limitations in forecasting routine weather events. To overcome these challenges, we propose a novel model, DC-CNN-BiLSTM, which integrates a dilation causal convolutional neural network (DC-CNN) with a Bidirectional Long Short-Term Memory (BiLSTM) network. The DC-CNN component, by fusing causal and dilated convolutions, extracts multi-scale spatial features from time series data. In parallel, the BiLSTM module leverages bidirectional memory cells to capture long-term temporal dependencies. This integrated approach effectively links localized meteorological inputs with broader hydrological responses. Experimental evaluation demonstrates that the DC-CNN-BiLSTM model significantly outperforms traditional models. Specifically, the model improves the Root Mean Square Error (RMSE) by 9.05% compared to ConvLSTM and by 32.3% compared to ConvGRU, particularly in forecasting medium- to long-term precipitation. In conclusion, our results validate the benefits of incorporating advanced spatio-temporal feature extraction techniques for precipitation forecasting, ultimately improving disaster preparedness and resource management.

1. Introduction

Precipitation prediction stands at the intersection of meteorology, climate science, and computational modeling, with profound implications for sustainable development and disaster resilience. In agriculture, precise precipitation forecasts enable optimized irrigation scheduling, reducing water waste by up to 30% while ensuring crop hydration [1]. Within the hydrological cycle, precipitation regulates the delicate balance between freshwater availability and saline ecosystems, directly impacting global water resource management [2]. Accurate forecasting is indispensable for urban water supply planning, industrial operations, and ecological conservation [3,4]. Furthermore, reliable predictions are critical for mitigating extreme weather events, such as floods and droughts, which have increased in frequency and intensity due to climate change [5,6].
Traditional forecasting methods primarily rely on meteorological observation techniques and numerical weather prediction, with the latter playing a central role in precipitation forecasting. Meteorological observation methods use data from observation stations—such as temperature, humidity, barometric pressure, and wind speed and direction—combined with historical data and empirical models to make precipitation forecasts. In contrast, numerical weather prediction relies on atmospheric dynamics and thermodynamics equations to simulate atmospheric evolution. This approach involves dividing the atmosphere into a grid and solving the fundamental equations of motion, heat, and moisture conservation at each grid point. Pu proposed the fundamentals of the numerical weather prediction (NWP) model [7], including the numerical framework, methodology, and physical parameterization used to simulate atmospheric dynamics; Vasić proposed to further improve the accuracy of the NWP model and satellite precipitation estimation by incorporating the meteorological data from radar observations [8]. Despite their widespread use, these methods face significant challenges. NWP models, while theoretically robust, require extensive computational resources and time, limiting their effectiveness for short-term, high-resolution predictions [8]. Additionally, flash flood forecasting remains constrained by insufficient temporal resolution and data integration capabilities [9,10]. These limitations underscore the need for more advanced, efficient prediction frameworks.
The integration of artificial intelligence (AI) has revolutionized precipitation forecasting, particularly through the application of deep learning architectures such as convolutional neural networks (CNNs) and Long Short-Term Memory (LSTM) networks. CNNs excel at capturing spatial patterns in precipitation data, while LSTMs model temporal dependencies, making them ideal for spatio-temporal prediction tasks [11,12]. Notable advancements include the following: Zhang [13] developed a CNN-based framework incorporating spatio-temporal priors, achieving superior performance in short-term rainstorm prediction; Ishida [14] introduced CNNsLSTM, a deep learning model for rainfall-runoff modeling, demonstrating high accuracy in hourly river flow estimation.
Zheng [15] proposed a precipitation nowcasting model based on spatio-temporal co-attention, which significantly improved the prediction accuracy of short-term heavy precipitation events by dynamically fusing spatial features and temporal dependencies in radar echo sequences. And Chen [16] proposed the STAT-LSTM model, which significantly improved the ability of SPEI drought prediction to capture complex climate dynamics by fusing multivariate spatio-temporal features and an adaptive feature aggregation mechanism.
Despite these advancements, challenges persist. Long-term spatio-temporal predictions often suffer from catastrophic forgetting, where models lose previously learned information, leading to suboptimal training outcomes [17,18,19]. Additionally, model selection and hyperparameter tuning significantly impact prediction accuracy, necessitating further research [20]. These models usually face the limitation of spatio-temporal feature fusion and the difficulty of modeling complex climate phenomena.
To address these limitations, we researched how to combine the temporal and spatial dependence of precipitation to improve the accuracy of precipitation prediction. We propose a novel deep learning model, DC-CNN-BiLSTM, which incorporates the spatial-temporal dependencies of precipitation data by combining a dilatation causal convolutional neural network with a Bidirectional LSTM model and a multi-head attention mechanism, resulting in the DC-CNN-BiLSTM model. Specifically, the causal convolution efficiently captures spatial dependencies between regions, while the BiLSTM captures temporal features. With this innovation, our model demonstrates greater advantages in dealing with complex spatio-temporal dependency problems and greater predictional accuracy of precipitation events.
The contributions of this paper can be summarized as follows. The proposed DC-CNN-BiLSTM framework uniquely integrates a DC-CNN and BiLSTM to achieve efficient feature extraction while effectively capturing spatio-temporal dependencies. Furthermore, by restricting predictions to past information, the model incorporates causal inference principles that prevent data leakage and ensure physically consistent predictions. The experimental results further show that the proposed method is superior to traditional statistical and deep learning methods. For example, compared to traditional ConvLSTM and ConvGRU models, DC-CNN-BiLSTM improves prediction accuracy by 9.05% and 32.3% (RMSE). In particular, its effectiveness in improving prediction accuracy is highlighted in the scenario of precipitation over medium and long periods of time.
The proposed DC-CNN-BiLSTM model represents a significant step forward in precipitation prediction, addressing critical limitations in traditional and AI-driven approaches. By leveraging causal inference and advanced deep learning techniques, this framework offers improved accuracy, efficiency, and applicability across agricultural, hydrological, and disaster management domains. Future research will focus on fusing this model with multi-source meteorological data to further enhance operational forecasting capabilities.

2. Methodology

2.1. BiLSTM Neural Network

Long Short-Term Memory (LSTM) is a variant of Recurrent Neural Networks (RNNs) [21] widely used in deep learning. The Bidirectional Long Short-Term Memory (BiLSTM) neural network is an extension of traditional LSTM, designed to capture both past and future context in sequential data. LSTM networks are particularly effective at processing long sequences by addressing the vanishing gradient problem inherent in traditional RNNs. BiLSTM enhances this by using two LSTM layers: one processes the input sequence in chronological order, while the other processes it in reverse order. This bidirectional structure allows the model to capture dependencies from both directions, improving performance in tasks like precipitation prediction.
In precipitation forecasting, BiLSTM has shown superior accuracy in handling long-term dependencies compared to traditional LSTM models. By using both past and future data during training, BiLSTM better captures temporal relationships, leading to more accurate predictions. This architecture has been applied successfully in various sequence prediction tasks, including weather forecasting. Barrera-Animas [20] compared the applicability of modern machine learning methods to three variants of the LSTM network architecture for the task of predicting hourly rainfall. They also show that a Bidirectional LSTM network can be used as a rainfall forecast model.

2.2. Dilatation Causal Convolution

Causal convolution is performed on the basis of general convolution (CNN), in which the value of each time step of the output depends only on the values of the current and previous time steps of the input. In other words, the time step of the output cannot depend on the value of the future time step of the input, which ensures that the cause of the value of the current and previous time steps of the input produces the effect of the value of the current time step only, forming a causality constraint.
Dilated convolution expands the receptive field by inserting gaps (“holes”) between kernel elements, allowing the model to capture long-range dependencies without increasing the number of parameters. This technique preserves the output size and is effective in tasks that require broader context, such as image, audio, and time series processing.

2.3. Data Description

In this study, precipitation grid data from China from January to April 2020 were selected, and only the hourly precipitation data were included in the data. The reason we choose this particular time period (January to April) for analysis is that the late winter and early spring are often accompanied by the convergence of warm and cold air and it is easy to form special precipitation events, which are challenging for practical forecasting.
Table 1 contains a total of 120 days of data, and the data contain 24 h of data in tif format. The naming scheme of the data was year–month–day–hour–minute–second, with the year occupying 4 digits, month occupying 2 digits, day occupying 2 digits, and hour–minute–second occupying 4 digits. The spatial resolution was 1401 × 801, as shown in Table 1.
The data were pre-processed to organize them into a two-dimensional tensor, with dimension 0 being the daily 24 h data containing spatio-temporal sequence information and dimension 1 representing the number of days. The ratio of training set to validation set to test set was 8:1:1. The input channel was set to 1.
In order to show the spatial distribution of the original precipitation data more intuitively, we superimposed the precipitation data of a selected time step on the geographic base map of China in the form of pixels, as shown in Figure 1. In the figure, the color brightness is proportional to the precipitation intensity, and the greater the precipitation, the brighter the color of the region.

2.4. Definition

The idea for solving the time series forecasting problem was to mine the potential information and trends embedded in the data by modeling the historical data. The specifics can be described in mathematical form as Equation (1):
{ y t + 1 ,   y t + 2 ,   ,   y t + m } = f x t n ,   x t n + 1 ,   ,   x t 1 ,   x t
where x t is a one-dimensional vector containing features and y t + m is inferred value of current time t within time period m. The goal is to fit the function f as much as possible with all the data X{ x 1 ,   x 2 ,   ,   x t } and Y{ y 1 ,   y 2 ,   ,   y t } available in history, so as to infer the value of y at the next time from the historical data at any time, where x represents the known precipitation data and y represents the value of the precipitation data to be predicted.

2.5. Environment Configuration and Model Parameters

This experiment was built on a hardware configuration with 64 GB of RAM and a NVIDIA-RTX5000 GPU, using the Windows 10 operating system, torch version 2.1.0 + cuda121, and pycharm version 2023.2.4 for the experiment. Table 2 lists the configured hyperparameters.

2.6. DC-CNN-BiLSTM

Traditional LSTM models can only account for unidirectional temporal dependencies in time series, making them less effective for the complex and bidirectional nature of precipitation series. This limitation hampers accurate predictions. To address this, the DC-CNN-BiLSTM model integrates the feature extraction efficiency of the dilation causal convolutional neural network (DC-CNN) with the powerful sequence modeling capabilities of BiLSTM. Additionally, it leverages the unique advantages of a multi-attention mechanism to further enhance prediction accuracy. Due to the organic combination of these three components, the DC-CNN-BiLSTM model can more accurately capture complex dynamics in precipitation data.
Therefore, we propose the DC-CNN-BiLSTM model. As shown in Figure 2, the DC-CNN-BiLSTM hybrid neural network model constructed in this paper is a prediction model that takes a single feature time series variable as its input and a single variable as its output, and consists of an input layer, a causal convolutional layer, a multi-head attention (Ma) and BiLSTM layer, and an output layer. Among them, the dilatation causal convolution layer realizes the spatial dependence in the context of the input layer, and dilatation expands its vision. A flatten layer and dropout layer were added to optimize the model performance. The fully connected layer (FC-layer) is used to convert the output 2D feature map into a 1D vector. The error feature map is a graph generated using 10*10 errors of four different predictions. N is the batch size, T is the total time step, and E is the size of the embedding.
BiLSTM lacks the ability to learn cross-features and cannot better handle the multidimensional influences in precipitation prediction. And CNNs cannot handle the time series characteristics of precipitation prediction. Considering that CNNs’ strength lies in feature extraction and BiLSTM can effectively handle time series data [22], in order to take advantage of the two models, the two were combined to form a CNN-BiLSTM neural network model.

2.7. Model Details

Conventional convolutional neural networks typically capture spatial features by processing input data in parallel. However, when handling time series data or tasks with temporal dependencies, standard convolution may fall short in capturing causal relationships within the data. To address this issue, we introduced causal convolutions into neural networks, which incorporate causal dependencies along the time dimension. By introducing null coefficients, causal convolutions also account for dependencies in the spatial dimension. Furthermore, the Bidirectional Long Short-Term Memory network BiLSTM enhances the overall model by tightly integrating temporal and spatial relationships through its time-dependent structure.

2.7.1. DC-CNN for Spatial Dependencies

DC-CNNs, or dilatation causal convolutional neural networks, are obtained by adding an expansion coefficient to causal convolution, thereby increasing the sensory field of the convolutional kernel. The expansion coefficient determines the step size of the convolutional kernel jumps on the input data, allowing the network to efficiently capture long-distance dependencies in the input data without the need to increase the hierarchical depth. CNNs contain multiple feature extractors consisting of convolutional and pooling layers; pooling layers are not included in causal convolution. The general per neuron operation is to apply a linear transformation to the input, followed by a nonlinear function. Mathematically, this can be expressed as follows:
h w ,   b x = f i = 1 n w i x i + b
where h w , b x denotes the result of each neural network unit operation, i = 1 n w i x i , in which x i is the input matrix and w i is the weight matrix. A neural network model is formed when multiple units are combined and have a hierarchical structure. Specifically, the main functions of a convolutional neural network are to perform feature extraction on data and capture contextual sequences.
In time series tasks, standard convolution applies the kernel across the entire input, including past, present, and future time steps, which may unintentionally introduce future information. To preserve causality, masking is applied to remove connections from future time steps. Specifically, links from future positions are masked with -inf, causing the softmax operation to assign near-zero attention weights to those positions. Padding is then added to maintain the output size. This approach enables causal convolution, ensuring that the output at each time step depends only on the current and previous inputs, effectively capturing temporal and spatial dependencies. The input layer is a one-dimensional vector, as illustrated in Figure 3.
The dilation convolutional neural network expands the coverage of the convolution kernel to the historical data on this basis; i.e., the hollow dilation parameter is used during the movement of the convolution kernel to expand the coverage of the convolution kernel. In dilation convolution, a void coefficient (dilation rate) is inserted into the convolution kernel to expand the sensory field of the convolution kernel. The DC-CNN structure diagram after adding a dilation mechanism, when the input is a one-dimensional vector, is shown in Figure 4.
It can be observed that, after accounting for dilation convolution, the next layer of neurons in the causal neural network has a significantly enlarged field of view for sensing the historical information of the neurons in the previous layer. Through the mechanism of dilation, the current causal neural network can then expand the receptive field of view of the causal convolutional network by changing the size of the convolutional kernel (kernel_size) or increasing the dilation parameter while the number of hidden layers remains unchanged. Padding = (kernel_size-1)*dilation was chosen in this paper, and dilation was set to 2. Thus, the embedding process used causal and dilation convolution to capture the context and obtain the spatial relationship of the input data. Then, the context data were passed into the BiLSTM model using a tanh activation function.

2.7.2. Ma-BiLSTM for Contextual Temporal Dependencies

An Ma-BiLSTM is a Bidirectional Long Short-Term Memory network with the addition of multi-head attention mechanism. The multi-head attention mechanism was added in the class initialization stage, and the context data captured by causal convolution are imported. After adjusting the dimensional order of the processed data in the causal convolution, the weights continue to be adjusted by the multi-head attention mechanism.
LSTM is a variant of a Recurrent Neural Network which utilizes the logical control of gate units to realize the operation of data. BiLSTM adds three gating structures in the hidden layer h, which are a forget gate, input gate, and output gate, and also adds a hidden state [23]. f t , i t , and o t denote the values of the forget gate, input gate, and output gate at moment t, respectively; x t is the input value at the moment t; h t is the output value of the hidden layer at the current moment; c t is the state information at the current moment; and σ is the sigmoid activation function.
The forgetting gate determines the extent to which the cell state of the previous moment is retained in the current moment. Its output f t is represented by Equation (3):
f t = σ ( W f [ h t 1 ,   x t ] + b f )
The input gate determines the extent to which the input x t of the network at the current moment is saved to the cell state c t , and its output i t is represented by Equation (4):
i t = σ ( W i [ h t 1 ,   x t ] + b i )
In Equations (3) and (4), W f and W i denote the weight matrices of the forgetting gate and the input gate, respectively, and b f and b i denote the bias terms of the forgetting gate and the input gate, respectively. c ~ t is used to characterize the state of the current input unit and is represented as follows:
c ~ t = t a n ( W c [ h t 1 ,   x t ] + b c )
The memory unit is used to compute the cell state c t at the current moment, c t is represented by the sum of the cell state c t 1 at the previous moment multiplied by element by the forgetting gate f t , and the current input cell state c ~ t multiplied by element by the input gate i t :
c t = f t c t 1 + i t c ~ t
By Equation (6), the BiLSTM model’s memory c ~ t about the current moment and its long-term memory c t 1 are then associated in order to form a new memory c t .
In Equations (5) and (6), W c is the weight matrix of the memory cell and b c is the bias term of the memory cell.
The output gate controls the effect of long-term memory on the current output, which is processed by a sigmoid function to obtain the output o t of the unit:
o t = σ ( W o [ h t 1 ,   x t ] + b o )
The hidden layer output h t is jointly determined by the output gate and the new cell state:
f t = σ ( W f [ h t 1 ,   x t ] + b f )
In Equations (7) and (8), W o is the weight matrix of the output gate and b o is the bias term of the output gate.
Figure 5 shows the BiLSTM neural network structural model; its structure is similar to that of an LSTM. x t is the input to the network at the current moment, h t is the output of the hidden layer, and H t is the final feature expression obtained by splicing the positive-order and reverse-order outputs with t ∈ [1, n ]. H t is the eventually hidden state at the current moment, H t = [ h n , h 1 ] , and the output fully connects all hidden states. BiLSTM is optimized on the basis of LSTM; the BiLSTM neural network structure model is composed of two independent LSTM modules, and their input sequences are fed into the two LSTM neural networks in positive and reverse order for feature extraction, respectively. The parameters of the two LSTM neural networks in BiLSTM are independent of each other, and they only share the word-embedding word vector list. BiLSTM predicts precipitation using time series within time steps, rather than being validated against future predicted precipitation.
The resulting DC-CNN-BiLSTM model combines the strengths of the DC-CNN and the Ma-BiLSTM network, creating a hybrid model.

2.8. Loss Function

In the whole prediction model, the Mean Squared Error (MSE), Root Mean Squared Error (RMSE), and Mean Absolute Error (MAE) were selected to calculate the error loss. Loss is evaluated using MSE at Train Loss and Eval Loss, and RMSE at validation. Mean Squared Error refers to the expected value of the squared difference between the estimated value of a parameter and the true value of the parameter, which can evaluate the degree of change in the data and is denoted as L M S E . RMSE is the square root of the Mean Squared Error, which more intuitively represents the average error between the predicted values and the actual values, denoted as L R M S E . Compared to RMSE, MAE is relatively less sensitive to outliers, easier to interpret, and consistent with the units of the original data. In these two indicators, the smaller the value is, the more accurate the prediction of the model is. The formula is as follows.
L M A E = 1 n | y p y t |
L R M S E = ( 1 n y p y t 2 )
In Equations (9) and (10), y p and y t denote the normalized predicted and true values, respectively, and n is the total number of samples.

2.9. Model Building

In this study, a hybrid prediction model based on a DC-CNN, BiLSTM, and a multi-head attention mechanism was proposed. In the process of model construction, firstly, the multi-scale long-term dependence features of the input sequence were extracted through the dilated causal convolution structure of the DC-CNN module (the inflation factor was 20 to 22). Then, the high-level features output by the DC-CNN were input into the BiLSTM module to synchronously capture the history and future potential dynamic association through bidirectional time series modeling. In order to further enhance the utilization efficiency of key temporal information, a multi-head attention mechanism was introduced in front of the BiLSTM module to dynamically calculate the weight distribution of different time steps to focus on important signal segments.
The DC-CNN-BiLSTM model utilized Adam optimization with tanh and sigmoid activation functions, combined with Dropout regularization (0.5 rate), to mitigate overfitting. Training proceeded through sequential phases: forward propagation computed predictions and cross-entropy loss, followed by backpropagation deriving parameter gradients via automatic differentiation. These gradients then guided weight updates using Adam’s adaptive learning rates and momentum-based adjustments, iteratively minimizing prediction errors while maintaining model generalization through controlled parameter stochasticity and gradient-directed optimization. To evaluate the performance of the hybrid model, a random hierarchical partitioning strategy was adopted to divide the data into a training subset and a test subset. The training set was used to train the model, while the test set was used to evaluate the generalization ability of the model. During this process, the data were divided into k subsets, and the model was trained and tested on different combinations of these subsets to evaluate its performance more comprehensively. The final model performance was averaged over multiple runs to ensure stability and universality.
The performance of the model was evaluated using several key evaluation indicators. The Mean Absolute Error (MAE) was used to visually measure the magnitude of the prediction error. The Root Mean Square Error (RMSE) was chosen because of its sensitivity to large errors, as it can penalize large deviations between predicted values and actual values. The Mean Squared Error (MSE) was used as a loss function in training to calculate the average square difference between the predicted values and the actual values, providing a sense of the accuracy of the model. Finally, the coefficient of determination (R2) was used to quantify the extent to which the model explained the variance in the target variable. The higher the R2, the better the overall fit.
Ablation studies were conducted to evaluate the contribution of each module in the model. In this study, the DC-CNN and Ma-BiLSTM modules were removed one by one to evaluate their individual influences on the model performance. In the first variant, the DC-CNN module was removed and only BiLSTM was used to evaluate the model. In the second variant, the BiLSTM module was excluded, leaving DC-CNN and the attention mechanism. The same MAE and RMSE were used to measure the influence of each indicator on each module.

3. Results

3.1. Analysis of Results

RMSE (Equation (10)), Root Mean Square Error, and MAE (Equation (9)), Mean Absolute Error, were selected as evaluation indexes for precipitation prediction.
The model stops after the model reaches the epoch number. According to Train Loss and Eval Loss, used to judge the effects during training, if both decline, this indicates that the model is still learning and should continue to train; if the former declines and the latter rises, it indicates that the model is beginning to overfit and that it should be stopped and we should reduce the complexity of the model or increase the amount of data in the training set, and so on. The results are shown in Figure 6, a t1 = 10 performance evaluation plot, where RMSE is calculated for RMSE loss and MAE is calculated for MAE loss; the smaller the values, the more accurate the prediction. To avoid chance events, our experimental model results take the average of 10 experiments. MAE reaches 0.00610, RMSE reaches 0.00788, and the average of the two evaluation metrics, the model average loss, reaches 0.00699. We used the model to predict at different time steps, and the graphs and tables in the paper present data in which the model was operating at a time step of t1 = 10, unless otherwise indicated.
Figure 7 illustrates the fit of the predicted and true values of the model and the root mean square loss at different time steps for the different time steps. In the figure, PRE denotes the predicted value, OBS denotes the true value, N represents the number of predicted and true values, and R 2 denotes the square of Pearson’s correlation coefficient, which is used to assess the degree of correlation between the two values. NSE is a statistical metric used to assess the accuracy of the model’s predictions, where 1 indicates that the model’s predictions match the observed data exactly. The black 1:1 line in the figure indicates that the predictions ideally match the observations, while the red regression line indicates the fit of the actual model. The scatter points are all concentrated at the edge of the black line, and the black and red lines basically overlap, indicating a good fitting effect. By analyzing these assessment indicators and graphs, the predictive performance and reliability of the model can be visually assessed.
Figure 8 presents a temporal comparison between the predicted and ground-truth-standardized precipitation values across 10 consecutive time steps at four distinct spatial grid points (subplots a–d). The vertical axis denotes the time progression, while the horizontal axis displays normalized (non-dimensional) precipitation values, enabling the direct evaluation of the model’s sequential prediction consistency. The alignment of the prediction trajectories with the observational data across all grid points highlights the model’s temporal stability and its capability to replicate precipitation dynamics under varying temporal conditions.
We use a scatter plot to illustrate the model’s prediction performance across China. Figure 9 displays the absolute differences between the predicted and observed values, where the model shows generally accurate predictions. The maximum absolute error is 0.5487 mm/day. The forecast accuracy is higher in the northeast, central, and western regions of China, while the prediction performance is slightly lower in coastal areas.

3.2. Comparison Experiment

We compared the prediction performance of the proposed method with the following models in the case of time step t1 = 10, and the results are shown in Table 3.
The models compared are as follows: Conv2d is a two-dimensional convolutional operation (2D CNN) which is commonly used to process data with two-dimensional grids and image data and is capable of extracting features in a two-dimensional space, such as precipitation on a geolocation grid [24]. Conv3d uses a three-dimensional convolutional neural network (3D CNN) for precipitation prediction which captures features in both spatial and temporal dimensions, leading to the integrated modeling of spatio-temporal data; however, this computation is relatively more complex and less efficient [25]. ConvLSTM is a model that combines convolutional neural networks (CNNs) and Long Short-Term Memory networks (LSTMs) for processing spatio-temporal data, especially video data and meteorological data [19]. ConvGRU is a deep learning model that combines a convolutional neural network (CNN) and a gated recurrent unit (GRU). ConvGRU enables the joint modeling of spatio-temporal data by using convolutional operations instead of fully connected operations in traditional GRUs [26]. TrajGRU is based on a trajectory gated cyclic unit model and adopts the idea of dynamic migration to adaptively adjust the spatial location of the hidden state according to the spatial information of the input data to enhance its modeling capability for complex spatio-temporal dependencies [27].
In Table 3, our model performs well in MAE and RMSE loss metrics and achieves the best prediction performance, showing its advantages in the precipitation prediction task. It can be seen that the experimental performance of our model is higher than the other models; the model uses dilation causal convolution to process the time series data, which improves efficiency and is more conducive to dealing with the spatial correlation of the time series data, while the BiLSTM network further captures the temporal dependence of the data, and the combination of the two makes the model’s effect better than the other models. ConvGRU is second only to our model, which stems from its better feature extraction for spatially correlated time series data and its lower computational efficiency compared to BiLSTM. ConvLSTM does not show a very good performance, probably due to the long prediction time steps. Conv3d is the least effective, and although it is suitable for processing time and spatial data, its training effect is poor due to its large computational resource overhead and unsuitability for processing time series data.
To comprehensively evaluate the predictive performance of the model, we employed four basic metrics to quantify the behavior of the model: True Positive (TP), True Negative (TN), False Positive (FP), and False Negative (FN). We classified the prediction results, and then we could accurately judge the prediction accuracy of some regions. We take a 24 h precipitation greater than 50 mm as the classification basis and precipitation less than 50 mm as the positive class. We also selected 20,000 grid areas, and the prediction was accurate if the threshold error was less than 0.1 mm. In Figure 10, TP represents the number of samples in the positive class that are correctly classified as positive class; TN represents the number of samples in the negative class that were correctly classified as negative; FP is the number of examples in the negative class that were incorrectly classified as positive; and FN is the number of samples in the positive class that were incorrectly classified as negative.
As shown in Figure 10, with the DC-CNN-BiLSTM model, most of the positive class samples are correctly classified and correctly classified, and there are no negative class samples in the selected grid data. The accuracy is 98.735%, calculated as in Equation (11):
A c c u r a c y =   ( T P + T N ) / ( T P + F N + T N + F P )
Meanwhile, we compared the performance of different models at different time steps, shown in Table 4. Our model has the best performance at longer time steps, while ConvGRU is only second to our model, mainly because ConvGRU also has a strong ability in capturing temporal correlation in time series, but it is not as good as our model, that combines a DC-CNN and a self-attention mechanism to extract complex spatial features. The performance of Conv2d and Conv3d is not much different, and Conv2d has the worst performance, but Conv2d performance increases with the increase in time steps, while Conv3d shows the trend of decreasing and then increasing; TrajGRU takes a long time to train and its performance is not good, and with the increase in time steps, the performance of TrajGRU becomes more and more poor. In Figure 11, we present a column chart using the RMSE data in Table 4, and the DC-CNN-BILSTM model shows significantly lower losses.

3.3. Ablation Experiment

The proposed model is formed by the combination of a dilation causal CNN and BiLSTM; the dilation causal CNN establishes the spatial characterization of the data and the BiLSTM extracts the temporal dependence. In order to validate the advantages of the combination of the two, the following ablation experiments were performed using the same hyperparameters:
  • DC-CNN: Only the spatially sensitive dilation causal convolution part of the model was used for prediction.
  • BiLSTM: Predictions were made using only the time-considering Bidirectional Long Short-Term Memory network part of the model.
Figure 12 shows the ablation experiments for the model across two time steps. The RMSE values of the models are higher than their MAE values. Meanwhile, DC-CNN-BiLSTM_RMSE and CNNBiLSTM_MAE are the losses computed using the RMSE and MAE for the overall DC-CNN-BiLSTM model. The others are similar, covering Train Loss and Eval Loss using the MSE loss method.
Table 5 and Figure 13 also take ten experiments using our model—with MAE and RMSE averages of 0.00610 and 0.00788, respectively—and, in the figure, the CNN has the highest loss, with an average of 0.47764, while the loss of the BiLSTM only reaches 0.33841. Also from Figure 13, it can be concluded that our model outperforms the two models individually at all different time steps. Thus, It can be shown that, by combining the two modules, our model works better than they do alone.
In addition, as shown in Table 6, we conducted ablation experiments at different time steps, with RMSE and MAE as the primary evaluation metrics. The Loss of our model at time step t1 = 5 is significantly higher compared to t1 = 10, indicating poorer prediction performance at shorter time steps. This is consistent with the focus of this study on long-term prediction, where accuracy improves with longer time horizons. Conversely, the performance of the single CNN and single Ma-BiLSTM models deteriorates as the time steps increase. The CNN model excels at capturing local spatio-temporal features but struggles with global spatio-temporal relationships in long time series. Although the BiLSTM model is effective at capturing temporal dependencies, it may fall short in handling complex spatial features, making it challenging to achieve accurate long-term predictions.

4. Discussion

Understanding the intrinsic spatio-temporal characteristics of precipitation is essential for the accurate prediction model in this paper. The spatio-temporal distribution characteristics of precipitation in this paper are essentially driven by complex meteorological processes, which means that the model must capture both their long-range dependence over time and cooperative variation in space. Previous studies have shown that precipitation series exhibit weak but significant long-range dependence in the temporal dimension, which is particularly prominent in extreme precipitation events and is enhanced with an increase in spatial aggregation scale [28,29,30]. This temporal persistence may be related to low-frequency oscillations in atmospheric circulation, such as the ENSO cycle, or the inertia of regional water vapor transport. At the same time, the spatial dependence of daily precipitation has been widely confirmed [31].
The DC-CNN-BiLSTM model shows significant advantages. Firstly, by introducing a Deep Convolutional Network (DC-CNN), the model can extract rich spatial local features from the original precipitation data and effectively identify the spatial distribution structure of heavy precipitation regions. At the same time, combined with the time-modeling ability of BiLSTM, the model can capture long-distance time dependencies, so as to describe the precipitation evolution process more accurately.
Although DC-CNN-BiLSTM has some effect, the model has certain limitations and areas for future improvement. Firstly, the dataset used in this study is limited in terms of precipitation attributes, containing only basic meteorological data. It does not incorporate other potential influencing factors, such as barometric pressure, wind speed, and humidity, which may restrict further improvements in prediction accuracy. This approach may place excessive reliance on statistical features in the data rather than capturing the underlying physical mechanisms. Consequently, despite the model’s high performance, its predictive physical consistency and reliability may be compromised when applied to extreme climate conditions or unconventional scenarios. Future work could enhance the interpretability and reliability of the model by incorporating additional meteorological variables or integrating methods with physical constraints, such as Das [32], who employ a hybrid physics–AI framework analogous to PINNs, integrating atmospheric dynamics equations with observational radar/satellite data to improve extreme precipitation nowcasting accuracy, particularly for short-term, high-intensity rainfall events.
Furthermore, there are uncertainties associated with the model performance, particularly in situations involving data distribution shifts or extreme climate scenarios not well represented in the training data. These uncertainties stem from both data limitations and the inherent complexity of atmospheric systems, which are not fully captured by data-driven models.
In addition, the performance of the model when dealing with high-frequency and low-frequency precipitation events also needs to be optimized. The current model still has a certain deviation in the prediction of extreme weather events, because too-low or too-high precipitation values for extreme weather will affect the overall precipitation prediction, and very few weather data make the prediction of extreme weather inaccurate. This is also part of the reason why the model’s RMSE values are generally higher than its MAE results. The model is also susceptible to issues such as gradient disappearance and gradient explosion.

5. Conclusions

In this paper, we introduced the DC-CNN-BiLSTM model, a deep learning-based approach designed to overcome the limitations of other deep learning-based precipitation prediction methods, such as ConvLSTM and ConvGRU, in understanding regional meteorological dynamics and forecasting weather events. By integrating a DC-CNN and BiLSTM, the model effectively captures both spatial and temporal dependencies, significantly enhancing precipitation forecasting accuracy. Our experimental results demonstrate that the proposed framework outperforms traditional methods, maintaining stable predictive performance and reducing forecast errors, particularly in medium–long-time-period precipitation scenarios. These findings underscore the model’s innovative contribution to advancing deep learning applications in meteorology.
Future research should focus on expanding the model’s capability by incorporating additional precipitation attributes to construct a more comprehensive multi-dimensional feature space. Enhancing the model’s structure will be crucial for better handling precipitation events with varying frequencies, while further optimizing the self-attention mechanism could improve prediction accuracy. Additionally, integrating advanced deep learning techniques, such as gradient penalty methods, can further enhance the model’s robustness and generalization ability.
Beyond technical advancements, this research holds significant implications for policy. More accurate precipitation forecasts can improve disaster preparedness, optimize water resource management, and support climate adaptation strategies. By strengthening predictive capabilities, this study contributes to the development of more resilient meteorological forecasting systems, aiding policymakers in mitigating the impact of some extreme weather events and enhancing long-term climate risk management.
Overall, the impressive performance of the DC-CNN-BiLSTM model in precipitation prediction highlights the significant potential of deep learning in weather forecasting. Future research should focus on further optimizing the model structure, incorporating additional meteorological data, and enhancing the model’s generalization ability and prediction accuracy. These efforts will contribute to developing a more comprehensive solution for addressing the complexities and dynamics of the ever-changing meteorological environment.

Author Contributions

Conceptualization, X.D. and P.S.; methodology, P.S.; software, P.S.; validation, J.L., Y.T., C.S., and S.L.; formal analysis, X.D. and C.S.; investigation, Y.T.; resources, X.D.; data curation, P.S.; writing—original draft preparation, P.S.; writing—review and editing, X.D.; visualization, J.L.; supervision, Y.T.; funding acquisition, X.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Guizhou Provincial Science and Technology Plan project [grant. Qian Kehe support PA [2025]001], National Natural Science Foundation of China [grant 42461057], Guizhou Provincial Basic Research Program (Natural Science) [grant. [2024]130], Qiankehe Platform KXJZ [2024]020], Qiankehe Basic ZK [2021]284], and the Fundamental Research Program of Shanxi Province [grant 202303021212164].

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors upon request.

Conflicts of Interest

Authors Xiaoqi Duan and Sheng Li are employed by Guizhou East Century Science Technology Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Kaplun, D.; Deka, S.; Bora, A. An intelligent agriculture management system for rainfall prediction and fruit health monitoring. Sci. Rep. 2024, 14, 512. [Google Scholar] [CrossRef] [PubMed]
  2. Duong, T.A.; Bui, M.D.; Rutschmann, P. A comparative study of three different models to predict monthly rainfall in Ca Mau, Vietnam. In Proceedings of the Wasserwirtschaft-Innovation aus Tradition, Gräz, Austria, 18–20 September 2018. [Google Scholar]
  3. Xiang, X.; Li, Q.; Khan, S.; Khalaf, O.I. Urban water resource management for sustainable environment planning using artificial intelligence techniques. Environ. Impact Assess. Rev. 2021, 86, 106515. [Google Scholar] [CrossRef]
  4. Wei, F.; Zhang, X.; Xu, J.; Bing, J.; Pan, G. Simulation of water resource allocation for sustainable urban development: An integrated optimization approach. J. Clean. Prod. 2020, 273, 122537. [Google Scholar] [CrossRef]
  5. Wu, M.; Wu, Z.; Ge, W.; Wang, H.; Shen, Y.; Jiang, M. Identification of sensitivity indicators of urban rainstorm flood disasters: A case study in China. J. Hydrol. 2021, 599, 126393. [Google Scholar] [CrossRef]
  6. Li, W.; Jiang, R.; Wu, H.; Xie, J.; Zhao, Y.; Song, Y.; Li, F. A System Dynamics Model of Urban Rainstorm and Flood Resilience to Achieve the Sustainable Development Goals. Sustain. Cities Soc. 2023, 96, 104631. [Google Scholar] [CrossRef]
  7. Pu, Z.; Kalnay, E. Numerical Weather Prediction Basics: Models, Numerical Methods, and Data Assimilation. In Handbook of Hydrometeorological Ensemble Forecasting; Springer: Berlin/Heidelberg, Germany, 2019. [Google Scholar]
  8. Vasić, S.; Lin, C.A.; Zawadzki, I.; Bousquet, O.; Chaumont, D. Evaluation of Precipitation from Numerical Weather Prediction Models and Satellites Using Values Retrieved from Radars. Mon. Weather. Rev. 2007, 135, 3750–3766. [Google Scholar] [CrossRef]
  9. Kelsch, M. Hydrometeorological Characteristics of Flash Floods; Springer: Berlin/Heidelberg, Germany, 2001. [Google Scholar]
  10. Hapuarachchi, H.A.P.; Wang, Q.J.; Pagano, T.C. A review of advances in flash flood forecasting. Hydrol. Process. 2011, 25, 2771–2784. [Google Scholar] [CrossRef]
  11. Zhang, Y.; Wallace, B. A Sensitivity Analysis of (and Practitioners’ Guide to) Convolutional Neural Networks for Sentence Classification. Computer science. 2015. [Google Scholar] [CrossRef]
  12. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  13. Zhang, T.; Liu, J.; Gao, C.; Wang, P.; Leng, L.; Xiao, Y. Prior-Guided gated convolutional networks for rainstorm forecasting. J. Hydrol. 2024, 633, 130962. [Google Scholar] [CrossRef]
  14. Ishida, K.; Ercan, A.; Nagasato, T.; Kiyama, M.; Amagasaki, M. Use of one-dimensional CNN for input data size reduction in LSTM for improved computational efficiency and accuracy in hourly rainfall-runoff modeling. J. Environ. Manag. 2024, 359, 120931. [Google Scholar] [CrossRef] [PubMed]
  15. Zheng, X.; Qin, H.; Wang, W.; Lei, W.; Shi, P. Research on precipitation nowcasting based on spatiotemporal cooperative attention. J. Earth Syst. Sci. 2025, 134, 48. [Google Scholar] [CrossRef]
  16. Chen, Y.; Wu, H.; Xie, N.; Liang, X.; Jiang, L.; Qiu, M.; Li, Y. STAT-LSTM: A multivariate spatiotemporal feature aggregation model for SPEI-based drought prediction. Earth Sci. Inform. 2025, 18, 1–15. [Google Scholar] [CrossRef]
  17. Ravuri, S.; Lenc, K.; Willson, M.; Kangin, D.; Lam, R.; Mirowski, P.; Fitzsimons, M.; Athanassiadou, M.; Kashem, S.; Madge, S.; et al. Skilful precipitation nowcasting using deep generative models of radar. Nature 2021, 597, 672–677. [Google Scholar] [CrossRef]
  18. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer using Shifted Windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 10012–10022. [Google Scholar]
  19. Shi, X.; Chen, Z.; Wang, H.; Yeung, D.-Y. Convolutional LSTM Network: A Machine Learning Approach for Precipitation Nowcasting. Adv. Neural Inf. Process. Syst. 2015, 28, 1–9. [Google Scholar]
  20. Barrera-Animas, A.Y.; Oyedele, L.O.; Bilal, M.; Akinosho, T.D.; Delgado, J.M.D.; Akanbi, L.A. Rainfall prediction: A comparative analysis of modern machine learning algorithms for time-series forecasting. Mach. Learn. Appl. 2022, 7, 100204. [Google Scholar] [CrossRef]
  21. Elman, J.L. Finding Structure in Time. Cogn. Cogn. Sci. 1990, 14, 179–211. [Google Scholar] [CrossRef]
  22. Li, T.; Hua, M.; Wu, X.U. A Hybrid CNN-LSTM Model for Forecasting Particulate Matter (PM2.5). IEEE Access 2020, 8, 26933–26940. [Google Scholar] [CrossRef]
  23. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L. Attention Is All You Need. Advances in neural information processing systems. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  24. Lecun, Y.; Bottou, L. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  25. Tran, D.; Bourdev, L.; Fergus, R.; Torresani, L.; Paluri, M. Learning Spatiotemporal Features with 3D Convolutional Networks. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015. [Google Scholar] [CrossRef]
  26. Ballas, N.; Yao, L.; Pal, C.; Courville, A.C. Delving Deeper into Convolutional Networks for Learning Video Representations. Computer Science. 2015. [Google Scholar] [CrossRef]
  27. Shi, X.; Gao, Z.; Lausen, L.; Wang, H.; Yeung, D.Y.; Wong, W.K.; Woo, W.C. Deep Learning for Precipitation Nowcasting: A Benchmark and A New Model. Adv. Neural Inf. Process. Syst. 2017, 30, 5618–5628. [Google Scholar]
  28. Iliopoulou, T.; Papalexiou, S.M.; Markonis, Y.; Koutsoyiannis, D. Revisiting long-range dependence in annual precipitation. J. Hydrol. 2018, 556, 891–900. [Google Scholar] [CrossRef]
  29. Iliopoulou, T.; Koutsoyiannis, D. Revealing hidden persistence in maximum rainfall records. Hydrol. Sci. J. 2019, 64, 1673–1689. [Google Scholar] [CrossRef]
  30. O’Connell, E.; O’Donnell, G.; Koutsoyiannis, D. On the spatial scale dependence of long-term persistence in global annual precipitation data and the Hurst Phenomenon. Water Resour. Resour. Res. 2023, 59, e2022WR033133. [Google Scholar] [CrossRef]
  31. Fan, C.; Yin, S.; Chen, D. Spatial correlations of daily precipitation over mainland China. Int. J. Climatol. 2021, 41, 6350–6365. [Google Scholar] [CrossRef]
  32. Das, P.; Posch, A.; Barber, N.; Hicks, M.; Duffy, K.; Vandal, T.; Singh, D.; van Werkhoven, K.; Ganguly, A.R. Hybrid physics-AI outperforms numerical weather prediction for extreme precipitation nowcasting. NPJ Clim. Atmos. Sci. 2024, 7, 282. [Google Scholar] [CrossRef]
Figure 1. The pixel-wise distribution of the original data over China. This figure shows the spatial distribution of the raw precipitation data in the geographical scope of China. The brighter the color of a region in the figure, the more precipitation the region received at the selected time (202001011300).
Figure 1. The pixel-wise distribution of the original data over China. This figure shows the spatial distribution of the raw precipitation data in the geographical scope of China. The brighter the color of a region in the figure, the more precipitation the region received at the selected time (202001011300).
Water 17 01381 g001
Figure 2. Structure of DC-CNN-BiLSTM prediction model.
Figure 2. Structure of DC-CNN-BiLSTM prediction model.
Water 17 01381 g002
Figure 3. Causal convolution of masks.
Figure 3. Causal convolution of masks.
Water 17 01381 g003
Figure 4. Dilation convolutional neural network.
Figure 4. Dilation convolutional neural network.
Water 17 01381 g004
Figure 5. BiLSTM neural network structural model.
Figure 5. BiLSTM neural network structural model.
Water 17 01381 g005
Figure 6. Model training loss and metrics curves over time steps: (a) loss metrics at time step t1 = 5 in training; (b) loss metrics at time step t1 = 10 in training.
Figure 6. Model training loss and metrics curves over time steps: (a) loss metrics at time step t1 = 5 in training; (b) loss metrics at time step t1 = 10 in training.
Water 17 01381 g006
Figure 7. The fitting of the predicted values and the real values of the model at different time steps, and the RMS loss at different time steps: (a) predicted values are fitted to true values at t1 = 5; (b) predicted values are fitted to true values at t1 = 10.
Figure 7. The fitting of the predicted values and the real values of the model at different time steps, and the RMS loss at different time steps: (a) predicted values are fitted to true values at t1 = 5; (b) predicted values are fitted to true values at t1 = 10.
Water 17 01381 g007
Figure 8. Comparison between predicted and ground-truth precipitation over 10 consecutive time steps: subplots (ad) correspond to standardized precipitation values at different spatial grid points.
Figure 8. Comparison between predicted and ground-truth precipitation over 10 consecutive time steps: subplots (ad) correspond to standardized precipitation values at different spatial grid points.
Water 17 01381 g008
Figure 9. A Pixel-wise visualization of prediction differences over China at a selected time step (t1 = 10). This figure illustrates the absolute differences between the predicted and observed values at a time step of t1 = 10. To emphasize the finer-scale discrepancies, the differences are color-coded as follows: pixels with zero difference are shown in red, while values within the range 0 < x < 1 are visualized using a multi-color gradient transitioning from orange to cyan.
Figure 9. A Pixel-wise visualization of prediction differences over China at a selected time step (t1 = 10). This figure illustrates the absolute differences between the predicted and observed values at a time step of t1 = 10. To emphasize the finer-scale discrepancies, the differences are color-coded as follows: pixels with zero difference are shown in red, while values within the range 0 < x < 1 are visualized using a multi-color gradient transitioning from orange to cyan.
Water 17 01381 g009
Figure 10. Various models’ confusion matrix regression result metrics.
Figure 10. Various models’ confusion matrix regression result metrics.
Water 17 01381 g010
Figure 11. Bar plots with respect to RMSE in Table 4.
Figure 11. Bar plots with respect to RMSE in Table 4.
Water 17 01381 g011
Figure 12. The ablation study, conducted to evaluate the impact of different model components on performance, illustrating the effect of removing the DC-CNN and the BiLSTM component: (a) ablation contrast experiment at t1 = 5; (b) ablation contrast experiment at t1 = 10.
Figure 12. The ablation study, conducted to evaluate the impact of different model components on performance, illustrating the effect of removing the DC-CNN and the BiLSTM component: (a) ablation contrast experiment at t1 = 5; (b) ablation contrast experiment at t1 = 10.
Water 17 01381 g012
Figure 13. Bar plots with respect to RMSE in Table 5.
Figure 13. Bar plots with respect to RMSE in Table 5.
Water 17 01381 g013
Table 1. Data table.
Table 1. Data table.
DataFormat Description
formattif
time span range1 January to 30 April (2020) 20200101–20200430
time resolution00:00–23:00
pixel resolution(0.05,0.05)
spatial resolution1401 × 801
data name (example)202001010000, 202001012300
Table 2. Hybrid model hyperparameters.
Table 2. Hybrid model hyperparameters.
HyperparametersNameValue
t0train time step92
t1prediction time step10
train_numtraining set16,000
validation_numvalidation set2000
test_numtest set2000
batch_sizebatch size500
lrinitial learning rate0.001
epochnumber of training sessions200
Table 3. Performance evaluation of different models at t1 = 10.
Table 3. Performance evaluation of different models at t1 = 10.
ModelMAERMSE
Conv2d0.261610.36020
Conv3d0.230410.35458
ConvLSTM0.234470.32911
ConvGRU0.073850.09664
TrajGRU0.168960.25399
DC-CNN-BiLSTM0.007880.00611
Table 4. Performance evaluation of different models at different time steps.
Table 4. Performance evaluation of different models at different time steps.
Time StepIndexConv2dConv3dConvLSTMConvGRUTrajGRUDC-CNN-BiLSTM
1RMSE0.340680.462620.224430.054820.191340.03234
MAE0.340680.462620.224430.054820.191340.03234
2RMSE0.375050.336360.233720.052390.187670.02006
MAE0.356660.316350.210710.048930.170180.01777
3RMSE0.351740.328810.223030.042530.180580.01161
MAE0.309830.285790.197340.036680.156190.00996
4RMSE0.343100.320730.230790.064760.166480.01108
MAE0.293280.270340.200710.054060.138110.00919
5RMSE0.341240.327150.253240.059300.182130.00924
MAE0.281270.265250.210650.047690.146220.00759
6RMSE0.345130.332690.250770.071800.191200.00811
MAE0.284700.267540.258380.055420.150050.00653
7RMSE0.357100.356910.248600.110930.187350.00836
MAE0.283400.284300.204590.084370.144380.00668
8RMSE0.360260.340050.277360.106650.215850.00797
MAE0.283690.265270.215980.076020.156230.00627
9RMSE0.370240.371170.294420.106440.245910.00871
MAE0.274060.282390.223220.076640.172430.00674
10RMSE0.385540.423340.329110.113710.279990.00788
MAE0.270670.304870.234470.082450.189800.00611
Table 5. Mean experimental evaluation at t1 = 10.
Table 5. Mean experimental evaluation at t1 = 10.
NumberDC-CNN-BiLSTMDC-CNNBiLSTM
MAERMSERMSERMSE
10.00510.00660.39330.3385
20.00700.00920.46450.3384
30.00510.00670.55480.3383
40.00640.00800.49380.3385
50.00440.00550.47210.3384
60.00550.00720.43960.3384
70.00490.00640.52830.3385
80.00600.00740.39620.3384
90.00860.01130.48330.3384
100.00770.01000.55050.3383
average0.006100.007880.477640.33841
Table 6. Performance evaluation of ablation models at different time steps.
Table 6. Performance evaluation of ablation models at different time steps.
Time StepDC-CNN-BiLSTMDC-CNNBiLSTM
RMSEMAERMSEMAERMSEMAE
10.032340.028590.296020.296020.218870.21887
20.032340.017770.289890.258550.233750.21074
30.011610.009960.300980.265790.226490.19867
40.011080.009190.300170.245670.230980.20089
50.008030.006590.353660.281090.257770.21122
60.010850.008590.381510.320830.261320.21047
70.009690.007700.399720.320940.258220.20655
80.010570.008230.475430.384160.282380.21673
90.008710.006740.421400.311590.304070.22294
100.007880.006110.409710.288380.338370.23600
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Shu, P.; Duan, X.; Shao, C.; Liu, J.; Tian, Y.; Li, S. Precipitation Spatio-Temporal Forecasting in China via DC-CNN-BiLSTM. Water 2025, 17, 1381. https://doi.org/10.3390/w17091381

AMA Style

Shu P, Duan X, Shao C, Liu J, Tian Y, Li S. Precipitation Spatio-Temporal Forecasting in China via DC-CNN-BiLSTM. Water. 2025; 17(9):1381. https://doi.org/10.3390/w17091381

Chicago/Turabian Style

Shu, Peng, Xiaoqi Duan, Chenming Shao, Jie Liu, Youliang Tian, and Sheng Li. 2025. "Precipitation Spatio-Temporal Forecasting in China via DC-CNN-BiLSTM" Water 17, no. 9: 1381. https://doi.org/10.3390/w17091381

APA Style

Shu, P., Duan, X., Shao, C., Liu, J., Tian, Y., & Li, S. (2025). Precipitation Spatio-Temporal Forecasting in China via DC-CNN-BiLSTM. Water, 17(9), 1381. https://doi.org/10.3390/w17091381

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop