Previous Article in Journal
Integrating Multi-Temporal Landsat and Sentinel Data for Enhanced Oil Palm Plantation Mapping and Age Estimation in Malaysia
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Large-Scale Flood Detection and Mapping in the Yangtze River Basin (2016–2021) Using Convolutional Neural Networks with Sentinel-1 SAR Images

1
College of Urban and Environmental Sciences, Hubei Normal University, Huangshi 435002, China
2
Key Laboratory of Digital Earth Science, Aerospace Information Research Institute, Chinese Academy of Sciences, Beijing 100094, China
3
Department of Environment and Society, Quinbey College of Natural Resources, Utah State University, Logan, UT 84322, USA
4
School of Geography, Development and Environment, The University of Arizona, Tucson, AZ 85719, USA
5
International Research Center of Big Data for Sustainable Development Goals, Beijing 100094, China
6
College of Ecological Engineering, Guizhou University of Engineering Science, Bijie 551700, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(16), 2909; https://doi.org/10.3390/rs17162909 (registering DOI)
Submission received: 11 June 2025 / Revised: 12 August 2025 / Accepted: 13 August 2025 / Published: 21 August 2025

Abstract

Synthetic Aperture Radar (SAR) technology offers unparalleled advantages by delivering high-quality images under all-weather conditions, enabling effective flood monitoring. This capability provides massive remote sensing data for flood mapping, while recent rapid advances in deep learning (DL) offer methodologies for large-scale flood mapping. However, the full potential of deep learning in large-scale flood monitoring utilizing remote sensing data remains largely untapped, necessitating further exploration of both data and methodologies. This paper presents an innovative approach that harnesses convolutional neural networks (CNNs) with Sentinel-1 SAR images for large-scale inundation detection and dynamic flood monitoring in the Yangtze River Basin (YRB). An efficient CNN model entitled FloodsNet was constructed based on multi-scale feature extraction and reuse. The study compiled 16 flood events comprising 32 Sentinel-1 images for CNN training, validation, inundation detection, and flood mapping. A semi-automatic inundation detection approach was developed to generate representative flood samples with labels, resulting in a total of 5296 labeled flood samples. The proposed model FloodsNet achieves 1–2% higher F1-score than the other five DL models on this dataset. Experimental inundation detection in the YRB from 2016 to 2021 and dynamic flood monitoring in the Dongting and Poyang Lakes corroborated the scheme’s outstanding performance through various validation procedures. This study marks the first application of deep learning with SAR images for large-scale flood monitoring in the YRB, providing a valuable reference for future research in flood disaster studies. This study explores the potential of SAR imagery and deep learning in large-scale flood monitoring across the Yangtze River Basin, providing a valuable reference for future research in flood disaster studies.

1. Introduction

In the context of global warming, the frequency of extreme weather-related disasters is escalating, resulting in significant human and economic consequences [1]. Among these disasters, floods stand out as the most devastating. Between 1995 and 2005, weather-related disasters affected an average of 205 million people annually, causing substantial casualties and damage [2]. The Yangtze River, Asia’s longest and the world’s third-longest river, holds paramount importance for China. It traverses diverse ecosystems, irrigates one-fifth of China’s land, and sustains one-third of the nation’s population. Historically, the region has been plagued by severe flooding, particularly during the East Asian monsoon season, primarily in June and July [3,4]. Notably, the Yangtze River Basin (YRB) experienced 24 major flood events caused by heavy rainstorms, ranking among China’s top ten natural disasters from 2005 to 2020, resulting in significant loss of life and property [5]. Given the YRB’s socio-economic significance and susceptibility to frequent floods, effective real-time flood monitoring is crucial, not only for the YRB but for all of China [6].
Satellite remote sensing plays a crucial role in flood detection by swiftly and accurately delineating inundated areas [7]. Flood monitoring primarily relies on two types of satellite data: optical images from sensors such as MODIS, Landsat, and Sentinel-2, and Synthetic Aperture Radar (SAR) images [8,9,10]. While optical sensors are valuable, their effectiveness can be hindered by heavy cloud cover associated with flood events, impeding the acquisition of precise surface information [11,12,13]. In contrast, SAR offers distinct advantages in flood monitoring, functioning under all-weather and all-day conditions, unaffected by cloud cover [11,14,15]. Moreover, SAR’s ability to capture variations in backscattering intensity across different land cover types makes it adept at identifying water bodies within its images. These capabilities have established SAR images as a cornerstone in real-time flood disaster monitoring [16,17,18].
Unlike optical imagery, SAR images require intricate pre-processing due to their imaging mechanisms and noise, which involve procedures such as orbit correction, radiometric calibration, and topographic corrections [18,19]. Various SAR-based flood detection methods have emerged, including threshold-based, region-growing, change detection, and deep learning approaches [20,21,22,23]. The threshold method, a swift and effective image segmentation technique, distinguishes floodwater from the background based on global or regional thresholds [24,25]. However, using a single unchanged threshold for the entire image may compromise accuracy. To address this limitation, subdividing the study area with distinct regional thresholds and constraints [7,26], with the Otsu method, a statistic-based threshold segmentation method, was widely employed in flood detection [27,28]. Moreover, region-growing and change detection methods, each with unique advantages and drawbacks, can be frequently found in relevant studies [29,30,31]. Region-growing excels at boundary segmentation, while change detection effectively rectifies issues such as mistaking mountain shadows for water. Nonetheless, the computational demands for the region-growing scheme are high, while the change detection approach is highly sensitive to the noise of SAR images [32,33]. In summary, the laborious pre-processing and extensive expert involvement in traditional flood detection methods significantly undermine their accuracy and efficiency, rendering them unsuitable for real-time flood monitoring.
To address the limitations of conventional flood monitoring and mapping techniques, deep learning methodologies have gained substantial prominence across diverse remote sensing disciplines, encompassing classification, target detection, risk assessment, and beyond [34,35]. In particular, Convolutional Neural Networks (CNNs), designed as end-to-end image processing models tailored for handling large-scale datasets and intricate classification tasks, have garnered significant attention in recent years. The noteworthy progress in this field includes the introduction of the unsupervised classification model “Felz CNN” for flood mapping in the Yangtze River region [36] as well as the comprehensive evaluation of various CNN models used for flood monitoring in the Poyang Lake region (such as HRNet, DenseNet, SegNet, ResNet, and DeepLab-v3+) [37]. It is noteworthy that global flood datasets have become available for extensive flood-related studies. The United Nations Satellite Centre (UNOSAT) has compiled a repository of over 200 flood events since 2007, resulting in the creation of the UNOSAT Flood Dataset, which leverages various satellite sensors. Additionally, Sen1floods11, a flood detection dataset tailored for deep learning applications, utilizes Sentinel-1 and Sentinel-2 images [38]. Numerous studies have scrutinized the performance of CNNs using these datasets [39,40,41]. In comparison to traditional methodologies, CNNs offer distinct advantages, including high automation, robust scalability, and superior efficiency and accuracy. Nonetheless, the utilization of deep learning technology in the realm of large-scale flood monitoring remains underexplored. For instance, while Reference [37] applied CNNs to flood monitoring in Poyang Lake, this approach has not been extended to large-scale monitoring across the Yangtze River Basin. Additionally, there is a current lack of publicly accessible datasets that can be directly utilized for basin-wide flood monitoring in the YRB.
In this study, we propose an efficient CNN model named FloodsNet, designed for large-scale flood detection and mapping. This model was deployed to perform inundation detection and flood mapping from 2016 to 2021 within the YRB, with the key highlights of the study summarized as follows:
(1)
An accurate flood dataset employing a semi-automatic approach with region thresholding and manual interpretation involved was generated.
(2)
An efficient CNN model, FloodsNet, that harnesses feature reuse and a spatial pyramid mechanism to enhance flood detection capabilities was proposed.
(3)
Inundation detection accuracy under various pre-processing strategies, including polarization, decibel conversion, and DEM adjustment, was systematically assessed.
(4)
Flood detections with the FloodsNet in the Yangtze River Basin spanning from 2016 to 2021, and dynamic monitoring for the floods in Dongting Lake in 2017 and Poyang Lake in 2020 were implemented.
The paper’s structure is organized as follows. Section 2 introduces the study area, data sources, and the proposed methodology. Section 3 outlines the comparative experiments conducted and presents their corresponding results. In Section 4, a synthesis of the experimental findings is provided, along with comprehensive discussions. Finally, Section 5 offers concluding remarks summarizing the key insights and implications of the study.

2. Materials and Methods

2.1. Study Area

The Yangtze River Basin (YRB), depicted in Figure 1, stands as China’s largest basin, covering an expansive drainage area of approximately 1.8 million square kilometers, equivalent to 18.8% of China’s total territory. Originating from the Tanggula Mountains, the Yangtze River stretches an impressive length of 6300 km, traversing 11 provinces. The YRB’s intricate topography showcases a network of tributaries cascading through plateaus, mountains, hills, and plains, before converging in eight other provinces, resulting in its river channels winding through a total of 19 provinces.
The YRB features diverse landforms, including the elevated Qinghai-Tibet Plateau, the Sichuan Basin, the plains within the middle and lower reaches of the Yangtze River, and the hilly regions of Jiangnan, among others. Renowned for its robust agricultural and economic activities, the basin plays a pivotal role in China’s overall grain output and Gross Domestic Product (GDP), contributing approximately 32% and 35% of the nation’s totals, respectively.
The region experiences the influence of a monsoon climate, resulting in uneven temporal and spatial distribution of rainfall. The YRB’s rainy season begins in April and extends through September. Particularly, the middle and lower reaches witness concentrated and sustained rainfall, known as “plum rain,” occurring predominantly from June to July. This climatic pattern often leads to flooding in flat, lower-lying areas, notably in the Dongting and Poyang Lake regions.
Characterized by its advanced agricultural and economic sectors, coupled with recurring flood events, the YRB serves as an exemplary region for in-depth studies related to flood disasters.

2.2. Data

In the preliminary assessment of satellite image availability for our study area, it became apparent that optical satellite images collected during flooding events were significantly obscured by cloud cover. Therefore, this study focused on the period from 2016 to 2021, specifically targeting flooding events within the Yangtze River Basin (YRB) as documented in the annual hydrological reports.
To meet the data requirements of this study, we utilized Sentinel-1, a C-band Synthetic Aperture Radar (SAR) satellite launched by the European Space Agency (ESA) in 2014. The YRB presented challenges due to cloud-covered optical images during flooding, but Sentinel-1’s SAR data proved indispensable. For this analysis, we acquired the Single Look Complex (SLC) and Ground Range Detected (GRD) products obtained in the interferometric wide swath mode (IW) from Sentinel-1. Given that flood detection relies on backscatter intensity information, the GRD data was chosen as the primary data source with an azimuthal resolution of 20 m. Additionally, we integrated a 12.5-m Digital Elevation Model (DEM) generated by ALOS-PALSAR as supplementary data, to assess the performance of our Convolutional Neural Network (CNN) model.
Table 1 presents detailed information on the image data utilized in this study. Each flood event considered in the analysis was represented by two images acquired during the flooding period and a non-flooding period for comparative purposes. These images were distinctly labeled based on their intended use for either training, testing, or application. The criteria for dividing the image data into training and testing datasets will be elucidated in Section 2.3.2.
In summary, this study involved 14 images capturing 7 flood events for both training and testing, 2 images from 1 flood event reserved exclusively for testing, and 16 images spanning the remaining 8 flood events, utilized solely for application purposes. An additional 16 images were simultaneously utilized for flood monitoring in Dongting Lake and Poyang Lake.

2.3. Methodology

The workflow of this study, depicted in Figure 2, comprises three fundamental components, each of which is elaborated upon in subsequent sections. The methodology includes:
  • Data Preprocessing: Prior to employing the threshold method for inundation area extraction, essential preprocessing steps were conducted. The significance and impact of these processes will be elucidated in the subsequent section.
  • Flood Dataset Generation: Deep learning methodologies critically depend on the quality and quantity of available datasets. To address this requirement, we introduced a semi-automatic approach that combined the global threshold method with regional thresholding, facilitating the creation of comprehensive flood datasets specific to the YRB.
  • CNN Model Development and Flood Detection and Dynamic Monitoring: This study involved a comparative evaluation, contrasting our proposed FloodsNet with a variety of classic CNN models. Subsequently, these models were applied in large-scale flood detection across the Yangtze River Basin, spanning the years 2016 to 2021. Additionally, our investigation encompassed a dynamic flood monitoring aspect, focusing on significant flood-prone regions such as Dongting Lake in 2017 and Poyang Lake in 2020 during their respective flood seasons.

2.3.1. Data Preprocessing

Given the inherent complexities of SAR imaging and the presence of substantial image noise, a series of preprocessing steps are indispensable. In accordance with previous studies [5,19] and Snap 8.0 software guidelines, we implemented six essential preprocessing algorithms:
(1)
Orbit Correction: This involved updating satellite orbit status information within the metadata file.
(2)
Thermal Noise Removal: Our objective was to eliminate noise originating from the SAR satellite system, particularly thermal noise.
(3)
Radiometric Calibration: Intensity data underwent systematic conversion into backscatter coefficient data, thereby enhancing the accuracy of our analysis.
(4)
Speckle Filtering: A crucial step dedicated to removing random speckle noise arising from radar echoes.
(5)
Terrain Corrections: Rectification of distortions induced by factors such as foreshortening, layover, or shadowing effects through the utilization of DEM.
(6)
Decibel Conversion: Converting the radar backscatter values from linear scale to dB scale: We performed the decibel conversion using logarithmic functions to enhance visualization and interpretation. The formula is dB = 10 × log10(P), where P is the intensity of radar echo.
These measures collectively ensure the preparedness of our SAR data for further processing and analysis, aligning with established best practices in remote sensing.

2.3.2. Label Annotation and Flood Dataset

To ensure accurate labeling, we adopted a semi-automatic approach for extracting inundated areas. Initially, we utilized a water index (WI) method based on Sentinel-1 data for image segmentation, employing the following formula:
W I = ln 10 × V H × V V 8
Here, VH and VV denote the two polarization bands after preprocessing. Subsequently, we applied a global threshold ranging from 0.3 to 0.4 to roughly segment 16 images corresponding to 8 flood events, taking into account image disparities. Concurrently, specific regions were earmarked as training and test datasets to optimize the efficacy of the deep learning model (further details on the selection strategy will be provided in the subsequent section).
It is pertinent to acknowledge that the similarity in backscattering coefficients between water and mountain shadows can lead to misclassification. To mitigate this, we utilized auxiliary DEM data and the region threshold method to refine the initial segmentation results. In undisturbed mountainous regions, we observed that applying a water index threshold of 0.15–0.2 alongside a slope threshold of 10° effectively rectified the preliminary segmentation outcomes.
Subsequently, the segmented results from the preceding step underwent manual annotation. During manual labeling, we referred to the Sentinel-1 images and water index to modify the previous segmentation results. Significant enhancements are evident in the outcomes, as illustrated in Figure 3, particularly with the incorporation of the global threshold method. The spatial distribution of the 16 flood events spanning from 2016 to 2021 in the YRB, derived from the annual hydrological report, is delineated in Figure 4.
Figure 4 also presents the distribution of the training and test data, encompassing various land cover types such as mountainous regions, hilly terrain, lakes, rivers, towns, and cultivated land. This diversity ensures the model’s ability to generalize and provides objectivity in evaluation. The selection of input channels holds significance in applying deep learning to remote sensing. As DEM was utilized in the threshold method during labeling, it was included as one of the channels for model training. To address the urgency of flood detection tasks, we retained the original VH and VV bands (Sentinel-1 GRD product data), designated as VH_ORI and VV_ORI, as input channels. Additionally, preprocessed VH and VV bands, labeled VH_PRE and VV_PRE, respectively, were integrated, bringing the total input channels to five.
The training dataset, derived from 14 images capturing 7 flood events, was partitioned into 256 × 256 tiles, resulting in a total of 5296 images. The test dataset was divided into two segments: test dataset 1, extracted from the training images (as depicted in Figure 4b–e), and test dataset 2, sourced exclusively from images designated for testing (as illustrated in Figure 4f). Test dataset 1 shares the same imaging conditions as the training data, while test dataset 2 is derived from a completely unfamiliar image. These datasets comprised a combined total of 13 images in test dataset 1 and 14 images in test dataset 2, each containing 30002–50002 pixels. Their primary purpose was to evaluate the generalization capability and robustness of the model.
During the testing and application phase, we utilized a sliding window algorithm with a window size of 256 and a step size of 128. Predicted results were generated within the middle quarter of the prediction window (128 × 128). Subsequently, all window predictions were aggregated to form the complete image.

2.3.3. The Proposed Deep Learning (DL) Model FloodsNet for Flood Mapping

In this study, we developed an efficient flood detection model, referred to as FloodsNet, outlined in Figure 5, aimed at extracting multi-scale features and facilitating the reuse of multi-level features for flood mapping. FloodsNet adopts the UNet architecture [42] as its base, with values such as 2562, 1282, and 642 representing the size of the feature map. The number of channels in all layers is reduced to 128 to minimize the model’s parameter count. The model is structured into down-sampling and up-sampling layers. The down-sampling layer comprises five convolutional blocks, each integrating a ResBlock [43] and a max-pooling layer. The ResBlock consists of three convolutional layers and incorporates a shortcut connection. To augment the model’s receptive field and enable the concurrent extraction of multi-scale features, we introduced two Atrous Spatial Pyramid Pooling (ASPP) layers in the last two convolutional blocks. At each up-sampling layer, we utilize the Skip Connection (SC) structure to merge features between adjacent layers, thereby enhancing feature reuse efficiency.
Atrous Spatial Pyramid Pooling (ASPP)
Figure 6a–c illustrate the operational mechanism of atrous convolution with dilated rates of 1, 2, and 3, respectively, enabling a larger receptive field than traditional convolution to capture features more effectively. The original ASPP employed dilated rates of 6, 12, and 18 to acquire a larger receptive field, which is beneficial for segmenting objects of different sizes.
However, flood segmentation prioritizes obtaining multi-scale detailed features rather than global features through a larger receptive field. Unlike natural images where the same object may exhibit features of varying sizes, floods in remote sensing images often cover a substantial area, rendering extensive dilated rates unnecessary. Figure 6d exhibits the improved structure of ASPP in FloodsNet, where the original dilated rates of 6, 12, and 18 were replaced with rates of 1, 2, and 3, respectively. Simultaneously, feature extraction was performed using an average pooling kernel with a size of 2 and a stride of 1. Inspired by ResNet, we aggregate convolution and pooling results together and add them with input feature maps to enhance the model’s multi-scale feature extraction and reuse capabilities. The ASPP structure can be expressed as:
F A S P P = F O + C o n c a t ( A C 1 ,     A C 2 ,   A C 3 ,   P 2 ) F o
where FO denotes the original feature maps of the input. AC1, AC2, AC3, P2 denote the atrous convolution with the dilated rates of 1, 2, 3 and the average pooling with the kernel size of 2, respectively. Concat denotes the function that aggregates the four feature maps.
Skip Connection (SC)
UNet’s architecture stands out for its ability to integrate features from both encoding and decoding layers, effectively addressing the challenge of losing edge and detail features. However, while this design leverages shallow features, it may not fully capture the semantic richness offered by deeper features. To bridge this gap, as illustrated in Figure 7, we introduced the SC structure, strategically employed to enhance deep semantic features during the decoding phase.
As illustrated in Figure 7, it can be observed that deep semantic features from the decoding phase are merged with shallow features from the encoding phase. The resultant integration is then combined with resampled deep features from the preceding layer, facilitating effective reuse of deep features. The SC structure can be expressed as:
F S C = C o n c a t D C 3 + C 3 + R 2
where DC3 denotes the deconvolution with kernel size of 3. C3 denotes the encoding feature maps of the corresponding encoding layer, that is, the shallow features. R2 denotes the feature map of up-sampling at the decoding layer, that is, the deep features. Concat denotes the function that aggregates the two feature maps.
The Cutting-Edge Models Adopted for Comparisons
In this section, we will revise and enhance the description of five well-established CNN models adopted in present study for comparisons in the comprehensive assessment of inundation detection and flood mapping in the YRB with our proposed FloodsNet. Table 2 presents the seminal literature and key characteristics of each model.

2.3.4. Evaluation Metrics and Experimental Parameters

For the evaluation of flood detection, we utilize a range of crucial metrics to gauge the model’s performance. These metrics are vital for binary classification and offer a comprehensive perspective on the outcomes. The key indicators employed include:
Overall Accuracy (OA): OA offers an overall assessment of the model’s performance. However, it may not be the most suitable metric when dealing with imbalanced positive and negative samples.
O A = T P + T N T P + T N + F P + F N
Precision: Precision measures the proportion of correctly predicted positive samples out of all the samples predicted as positive.
P r e c i s i o n = T P T P + F P
Recall: Recall quantifies the proportion of correctly predicted positive samples out of all the actual positive samples.
R e c a l l = T P T P + F N
F1_score: F1_score is a comprehensive metric that balances precision and recall, providing an overall measure of classification performance.
F 1 _ s c o r e = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
Kappa: Kappa assesses the agreement between observed accuracy and expected accuracy, accounting for random chance.
K a p p a = P 0 P e 1 P e
P 0 = O A
P e = ( T P + F P ) × ( T P + F N ) + ( F P + T N ) × ( F N + T N ) ( T P + T N + F P + F N ) 2
To compute these metrics, we analyze the model’s predictions, categorizing them as True Positives (TP), True Negatives (TN), False Positives (FP), and False Negatives (FN). These metrics collectively provide a comprehensive evaluation of our flood detection model, as elucidated in Table 3.
The experiments were conducted using the Tensor-flow framework on an NVIDIA GeForce RTX 2080Ti GPU. The specific parameter configurations employed in the experiments are detailed in Table 4. These parameters are determined through model training and are uniformly applied across all models.

3. Experiments and Results

3.1. Ablation Experiments

In our ablation experiments, we aimed to dissect the contributions of individual components within the flood monitoring model. Quantitative comparisons of CNN model performance were conducted using two distinct test datasets, and the results are summarized in Table 5.
The baseline model exhibited the lowest values across all evaluation indicators, except for precision. The inclusion of Resblock led to modest improvements in comprehensive evaluation metrics, specifically F1-score and Kappa. However, when ASPP and SC structures were integrated, substantial enhancements in all evaluation indicators were observed. Importantly, the ASPP structure demonstrated more significant improvement compared to the SC structure, highlighting the superior performance of a multi-scale feature structure over a feature reuse structure in flood mapping.
Our FloodsNet model, as introduced in this study, outperformed all the compared models across various metrics, particularly excelling in recall. It is worth noting that these models exhibited better performance with dataset 1 compared to dataset 2, primarily due to the markedly distinct imaging conditions in dataset 2, which encompass variations in factors such as atmospheric conditions, angles of incidence, system noise, and more, all of which were unfamiliar to the model.
To provide a detailed assessment of each model’s performance in flood detection, we conducted visual analysis on an image extracted from dataset 1, as depicted in Figure 8. The red boxes in the figure highlight areas where various model structures exhibited higher rates of false detections and misidentifications compared to FloodsNet.
Upon close examination, it becomes evident that the baseline model experienced significant issues with misidentification, such as incorrectly delineating water boundaries in the central section of Figure 8c and inaccurately detecting aquaculture regions in the lower-right part of the image.
The integration of ASPP (Figure 8e) and SC (Figure 8f) structures notably alleviated mis-detection, with ASPP exhibiting particularly significant improvement.
In contrast, our proposed FloodsNet model (Figure 8g) showcased minimal mis-detection, primarily limited to the aquaculture region. This visual assessment underscores the superior performance of FloodsNet in precise flood detection, with substantial reductions in mis-detection compared to the other structures.

3.2. Model Comparison Experiments

In our comparative model evaluation experiments, we conducted a quantitative evaluation of six distinct models using two separate datasets, as outlined in Table 6.
The findings indicate that FCN-8 displayed the weakest performance, possibly attributed to its direct resampling in the decoding layers without additional training. Among the remaining models, UNet, DeepLabv3, and DeepResUNet showcased similar results, with closely aligned F1-score and Kappa values. Particularly, UNet exhibited the highest recall among all six models.
However, SegNet, lacking feature information from the encoding layers, exhibited lower accuracy compared to the other three models. Notably, our proposed FloodsNet model, incorporating multi-scale features and reusing multi-layer features, achieved superior accuracy across all metrics. Similarly, these models demonstrated better performance with dataset 1 compared to dataset 2.
Figure 9 illustrates the inundation detection results generated by the six models using an image from dataset 1. The red boxes in the figure highlight areas where the various models exhibited higher rates of wrong detection and mis-detection compared to FloodsNet.
Notably, FCN-8 (Figure 9c) displayed a considerable mis-detection rate in small inundated areas, indicating limitations in handling small objects. UNet (Figure 9d) exhibited a similar mis-detection rate, alongside a substantial wrong detection rate. The remaining three models (Figure 9e–g) demonstrated higher mis-detection rates than wrong detection rates. Conversely, our proposed FloodsNet model (Figure 9h) showcased a notably low wrong detection rate, with fewer instances of mis-detection. Particularly, the model excelled in detecting large water areas, highlighting its superior performance.

3.3. Band Comparison Experiments

In our band comparison experiments, we aimed to explore the impact of different bands on the classification performance of FloodsNet. We evaluated 14 different band combinations, and the results, presented in Table 7, offer valuable insights.
The findings clearly indicate that utilizing the VH_ORI band as the sole input leads to the best performance, with VH polarization demonstrating superior accuracy compared to VV polarization. Surprisingly, the preprocessed bands VH_PRE and VV_PRE not only failed to enhance the model’s performance but also decreased its accuracy. Additionally, the incorporation of DEM had minimal effect on flood detection in our study. Furthermore, the introduction of additional inputs into the model did not yield positive contributions to its performance; rather, it seemed to introduce both information and noise, negatively impacting the classification performance. Moreover, the results from this section indicate that directly utilizing Sentinel-1 GRD data for flood mapping offers improved efficiency and accuracy.
Figure 10 illustrates examples of the results from the band combination experiment using an image from dataset 2. The lowest rates of wrong and mis-detected inundation areas were achieved when using only the VH_ORI band as input (Figure 10e). In contrast, introducing the VV polarization (Figure 10d,f,h,j) led to the emergence of numerous mis-detected inundation areas, highlighting the introduction of noise associated with VV polarization.
The inclusion of DEM in the inputs had minimal effect on the results, as demonstrated in Figure 10c,e,g,i. The mis-detected inundation areas were primarily concentrated along the boundaries of the inundated area, where SAR image backscattering coefficients exhibited similarities. These findings underscore the importance of selecting appropriate bands for flood detection to minimize noise and enhance classification accuracy.

3.4. Flood Monitoring Results

After conducting a comprehensive evaluation of our proposed model, which involved experimental comparisons with various popular models and different band combinations as inputs, we have determined that utilizing the VH_ORI band alone as input is the most effective approach for inundation detection and flood mapping in the YRB. The flood monitoring results, spanning from 2016 to 2021, are showcased in Figure 11.
These findings underscore significant flood events occurring within the YRB during the years 2016, 2017, and 2020. Particularly noteworthy are the floods observed in 2016 along the middle reaches of the YRB, in 2017 surrounding the Dongting Lake region, and in 2020 across the middle and lower reaches of the basin. Additionally, minor floods were recorded in some of the YRB’s tributaries during 2018, 2019, and 2021. The susceptibility of the middle and lower reaches of the YRB, as well as the Dongting Lake and Poyang Lake basin, to flooding can be attributed to various factors. These regions are characterized by persistent heavy rainfall from June to August, coupled with relatively flat terrain and the accumulation of sediments from upstream regions. Furthermore, increased anthropogenic activities contribute to their designation as primary flood-prone areas within the YRB.

4. Discussion

4.1. Polarization and DEM

According to [48], the VH polarization of SAR imagery provides a stronger echo signal in volume scattering and a weaker signal in specular reflection. This characteristic proves advantageous for flood monitoring, as it yields smoother and more homogeneous water surface images with reduced noise and variance between classes. Consequently, VH polarization is favored over VV polarization for flood detection [38,39]. Our own research, detailed in Section 3, further substantiates the superiority of VH polarization for inundation detection and flood mapping in the YRB. Inundation detection and flood mapping within the YRB, as depicted in Figure 12, demonstrated that VV band images exhibited significant noise, particularly in areas surrounding ships. This noise is likely attributed to interference caused by certain objects emitting electromagnetic waves at the same frequency. Notably, several studies have successfully used VV polarization for flood extraction [49,50]. Through our comparison with these studies, we found that VV polarization is particularly sensitive to water bodies, but it may result in misclassification in regions with dense vegetation.
Furthermore, SAR images acquired over mountainous terrain often exhibit shadows influenced by the surrounding topography, which can be erroneously interpreted as water due to their similar intensity [51,52]. Some studies have addressed this issue by utilizing the terrain slope to mask flood detection results, particularly in regions with significant topographical variations [9,11,53].
To delve deeper into the influence of DEM data in our investigation, we conducted statistical analyses on the DEM of both the training and test datasets, as depicted in Figure 13. Our analysis, depicted in Figure 13a, indicates that the average slope of the majority of images in the training dataset is less than 10 degrees. This observation corresponds with the distribution of floods depicted in Figure 4 and Figure 11, primarily within the middle and lower reaches of the YRB. Consequently, we infer that DEM has minimal impact on flood monitoring in these areas. Meanwhile, we added the mountain image in the training dataset, which can also avoid the model recognizing the mountain shadow as water.
It is important to note that our dataset comprises limited instances of flash floods detected in the YRB using Sentinel-1 between 2016 and 2021. For future endeavors, the inclusion of additional flash flood cases will be imperative to ensure the generalizability of the dataset.

4.2. Dynamic Monitoring of Floods in Typical Areas

To evaluate the efficacy and feasibility of our proposed methodology, we performed dynamic inundation detection and flood mapping for the Dongting Lake region in 2017 and the Poyang Lake region in 2020. The results are presented in Figure 14, Figure 15 and Figure 16.
In 2017, the flood season at Dongting Lake commenced in June, covering an initial area of approximately 630 km2. By July, most of the lake was inundated, reaching its peak around 5 July, with the flooded area nearly doubling in size. Subsequently, from 29 July to 10 August, the floodwaters gradually receded, resulting in a decrease in flood coverage. However, on 22 August, a new round of flooding was observed in the Dongting Lake region.
The results of flood monitoring for the Poyang Lake region in 2020 are illustrated in Figure 15. In June, the flood situation remained relatively stable. However, starting from 2 July, the inundated area began to escalate, reaching its peak on 14 July. Although the flood coverage started to decrease in August, it remained more extensive compared to June. It is noteworthy that both monitoring periods coincided with the flood season of the YRB, indicating that these events were attributed to flooding rather than variations in water levels between dry and wet periods.
By converting the aforementioned monitoring results into shapefile format and calculating their areas, we obtained the quantitative areal data presented in Figure 16. The dynamic flooding processes at Dongting Lake in 2017 and at Poyang Lake region in 2020 are statistically exhibited in the left and right of Figure 16 in terms of flooding area variations with time, respectively.

4.3. Potential and Limitations

In this research, we apply a deep learning model to conduct large-scale flood detection and mapping in the Yangtze River Basin (YRB). The all-weather, all-day imaging capabilities of Sentinel-1 SAR data provide crucial support for flood detection. The FloodsNet model proposed in this paper is applicable for future flood detection in the Yangtze River Basin, as demonstrated in the flood monitoring of Dongting Lake and Poyang Lake. Meanwhile, in small and medium-sized watersheds, the model also demonstrates good performance, as evidenced by the 2021 flood event in Figure 11d and the 2019 flood event in Figure 11f. It must be emphasized that the dataset and model in this study do not include any flash flood cases, which represents an area for improvement in future work. Although we use a semi-automated method to generate datasets more efficiently, the volume of the datasets remains limited. Looking ahead, we aim to make minor corrections to the flood detection results from our model and directly apply them to the production of flood datasets, making this work more meaningful.
While SAR satellites have an advantage in operational time compared to optical satellites, they are less accurate, especially when detecting specific features like wetlands and submerged vegetation. Recent studies [54] have attempted to augment flood mapping accuracy by integrating optical satellite data. In future work, it may be necessary to consider additional data sources and expand the scope of the current research in order to obtain multi-source remote sensing data that can support flood detection. Furthermore, integrating SAR and optical data to harness the advantages of both will be a challenge in future studies.

5. Conclusions

YRB stands out as one of China’s most flood-prone regions. Throughout this study, we have delineated a comprehensive approach for flood mapping, integrating both dataset production and algorithmic advancements, all rooted in deep learning methodologies.
Initially, we introduced a semi-automatic methodology tailored for generating flood datasets optimized for deep learning applications. This approach has substantially reduced both time and labor investments. Subsequently, we proposed a CNN model, FloodsNet, engineered for large-scale flood detection and mapping within the YRB. Leveraging multi-scale feature extraction and reuse, FloodsNet demonstrated its remarkable efficacy in our experiments. Our findings have underscored the dominance of VH polarization in flood detection, while also illustrating the minimal impact of DEM data on flood monitoring within the YRB. Importantly, FloodsNet emerged as the foremost model in our comparative analysis. Furthermore, the successful deployment of FloodsNet in flood detection and mapping within the YRB serves as a testament to its robust generalization capabilities.
In summary, this study not only reaffirms the effectiveness and superiority of deep learning methodologies in large-scale flood mapping but also emphasizes their vast potential in the field of flood monitoring.

Author Contributions

X.W. and Z.Z. designed this study. B.A., Z.L., and R.L. completed data collection and preprocessing. X.W. wrote this manuscript. Z.Z. provided suggestions and assistance for the experimental part. Z.Z., W.Z., and Q.C. revised and edited the manuscript. All authors have read and agreed to the published version of the manuscript without interest conflicts.

Funding

This research was jointly funded by the National Key R&D Program project (Grant No. 2023YFC3209102), and Major Science and Technology Projects (Grant Number: SKS-2022008) financed by the Ministry of Water Resources, China.

Data Availability Statement

The Sentinel 1 image used in the article can be downloaded from ESA through product ID (Link: https://dataspace.copernicus.eu, accessed on 11 July 2025). The flood monitoring results of the Yangtze River Basin from 2016 to 2021 and the dynamic flood monitoring results of Dongting Lake in 2017 and Poyang Lake in 2020 (SHP format) can be obtained by contacting the corresponding author.

Conflicts of Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

References

  1. IPCC Climate Change 2022—Mitigation of Climate Change—Working Group III; Cambridge University Press: Cambridge, UK, 2022; p. 1454.
  2. UNISDR; Centre for Research on the Epidemiology of Disaster CRED. The Human Cost of Weather Related Disasters; UNISDR: Geneva, Switzerland; CRED: Brussels, Belgium, 2015; Volume 12, pp. 1–17. [Google Scholar]
  3. Chen, Y.; Syvitski, J.P.M.; Gao, S.; Overeem, I.; Kettner, A.J. Socio-Economic Impacts on Flooding: A 4000-Year History of the Yellow River, China. Ambio 2012, 41, 682–698. [Google Scholar] [CrossRef]
  4. Wang, M.J.; Zheng, H.B.; Xie, X.; Fan, D.D.; Yang, S.Y.; Zhao, Q.H.; Wang, K. A 600-Year Flood History in the Yangtze River Drainage: Comparison Between a Subaqueous Delta and Historical Records. Chin. Sci. Bull. 2011, 56, 188–195. [Google Scholar] [CrossRef]
  5. Wu, X.; Zhang, Z.; Xiong, S.; Zhang, W.; Tang, J.; Li, Z.; An, B.; Li, R. A Near-Real-Time Flood Detection Method Based on Deep Learning and SAR Images. Remote Sens. 2023, 15, 2046. [Google Scholar] [CrossRef]
  6. Zhang, X.; Chan, N.W.; Pan, B.; Ge, X.; Yang, H. Mapping Flood by the Object-Based Method Using Backscattering Coefficient and Interference Coherence of Sentinel-1 Time Series. Sci. Total Environ. 2021, 794, 148388. [Google Scholar] [CrossRef]
  7. Shen, X.; Anagnostou, E.N.; Allen, G.H.; Robert Brakenridge, G.; Kettner, A.J. Near-Real-Time Non-Obstructed Flood Inundation Mapping Using Synthetic Aperture Radar. Remote Sens. Environ. 2019, 221, 302–315. [Google Scholar] [CrossRef]
  8. Feng, L.; Hu, C.; Chen, X.; Cai, X.; Tian, L.; Gan, W. Assessment of Inundation Changes of Poyang Lake Using MODIS Observations between 2000 and 2010. Remote Sens. Environ. 2012, 121, 80–92. [Google Scholar] [CrossRef]
  9. Gianinetto, M.; Villa, P.; Lechi, G. Postflood Damage Evaluation Using Landsat TM and ETM+ Data Integrated with DEM. IEEE Trans. Geosci. Remote Sens. 2006, 44, 236–243. [Google Scholar] [CrossRef]
  10. Huang, C.; Chen, Y.; Wu, J. Mapping Spatio-Temporal Flood Inundation Dynamics at Large Riverbasin Scale Using Time-Series Flow Data and MODIS Imagery. Int. J. Appl. Earth Obs. Geoinf. 2014, 26, 350–362. [Google Scholar] [CrossRef]
  11. Grimaldi, S.; Xu, J.; Li, Y.; Pauwels, V.R.N.; Walker, J.P. Flood Mapping under Vegetation Using Single SAR Acquisitions. Remote Sens. Environ. 2020, 237, 111582. [Google Scholar] [CrossRef]
  12. Martinis, S.; Rieke, C. Backscatter Analysis Using Multi-Temporal and Multi-Frequency SAR Data in the Context of Flood Mapping at River Saale, Germany. Remote Sens. 2015, 7, 7732–7752. [Google Scholar] [CrossRef]
  13. McCormack, T.; Campanyà, J.; Naughton, O. A Methodology for Mapping Annual Flood Extent Using Multi-Temporal Sentinel-1 Imagery. Remote Sens. Environ. 2022, 282, 113273. [Google Scholar] [CrossRef]
  14. Martinez, J.M.; Le Toan, T. Mapping of Flood Dynamics and Spatial Distribution of Vegetation in the Amazon Floodplain Using Multitemporal SAR Data. Remote Sens. Environ. 2007, 108, 209–223. [Google Scholar] [CrossRef]
  15. Zhao, J.; Pelich, R.; Hostache, R.; Matgen, P.; Cao, S.; Wagner, W.; Chini, M. Deriving Exclusion Maps from C-Band SAR Time-Series in Support of Floodwater Mapping. Remote Sens. Environ. 2021, 265, 112668. [Google Scholar] [CrossRef]
  16. Cohen, J.; Riihimäki, H.; Pulliainen, J.; Lemmetyinen, J.; Heilimo, J. Implications of Boreal Forest Stand Characteristics for X-Band SAR Flood Mapping Accuracy. Remote Sens. Environ. 2016, 186, 47–63. [Google Scholar] [CrossRef]
  17. Dasgupta, A.; Grimaldi, S.; Ramsankaran, R.A.A.J.; Pauwels, V.R.N.; Walker, J.P. Towards Operational SAR-Based Flood Mapping Using Neuro-Fuzzy Texture-Based Approaches. Remote Sens. Environ. 2018, 215, 313–329. [Google Scholar] [CrossRef]
  18. Ali, I.; Cao, S.; Naeimi, V.; Paulik, C.; Wagner, W. Methods to Remove the Border Noise from Sentinel-1 Synthetic Aperture Radar Data: Implications and Importance for Time-Series Analysis. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 777–786. [Google Scholar] [CrossRef]
  19. Tiwari, V.; Kumar, V.; Matin, M.A.; Thapa, A.; Ellenburg, W.L.; Gupta, N.; Thapa, S. Flood Inundation Mapping-Kerala 2018; Harnessing the Power of SAR, Automatic Threshold Detection Method and Google Earth Engine. PLoS ONE 2020, 15, e0237324. [Google Scholar] [CrossRef] [PubMed]
  20. Cian, F.; Marconcini, M.; Ceccato, P. Normalized Difference Flood Index for Rapid Flood Mapping: Taking Advantage of EO Big Data. Remote Sens. Environ. 2018, 209, 712–730. [Google Scholar] [CrossRef]
  21. Lê, T.T.; Froger, J.L.; Ho Tong Minh, D. Multiscale Framework for Rapid Change Analysis from SAR Image Time Series: Case Study of Flood Monitoring in the Central Coast Regions of Vietnam. Remote Sens. Environ. 2022, 269, 112837. [Google Scholar] [CrossRef]
  22. Li, Y.; Martinis, S.; Wieland, M. Urban Flood Mapping with an Active Self-Learning Convolutional Neural Network Based on TerraSAR-X Intensity and Interferometric Coherence. ISPRS J. Photogramm. Remote Sens. 2019, 152, 178–191. [Google Scholar] [CrossRef]
  23. Mason, D.C.; Speck, R.; Devereux, B.; Schumann, G.J.P.; Neal, J.C.; Bates, P.D. Flood Detection in Urban Areas Using TerraSAR-X. IEEE Trans. Geosci. Remote Sens. 2010, 48, 882–894. [Google Scholar] [CrossRef]
  24. Liang, J.; Liu, D. A Local Thresholding Approach to Flood Water Delineation Using Sentinel-1 SAR Imagery. ISPRS J. Photogramm. Remote Sens. 2020, 159, 53–62. [Google Scholar] [CrossRef]
  25. Wangchuk, S.; Bolch, T.; Robson, B.A. Monitoring Glacial Lake Outburst Flood Susceptibility Using Sentinel-1 SAR Data, Google Earth Engine, and Persistent Scatterer Interferometry. Remote Sens. Environ. 2022, 271, 112910. [Google Scholar] [CrossRef]
  26. Boni, G.; Ferraris, L.; Pulvirenti, L.; Squicciarino, G.; Pierdicca, N.; Candela, L.; Pisani, A.R.; Zoffoli, S.; Onori, R.; Proietti, C.; et al. A Prototype System for Flood Monitoring Based on Flood Forecast Combined with COSMO-SkyMed and Sentinel-1 Data. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2016, 9, 2794–2805. [Google Scholar] [CrossRef]
  27. Nakmuenwai, P.; Yamazaki, F.; Liu, W. Automated Extraction of Inundated Areas from Multi-Temporal Dual-Polarization Radarsat-2 Images of the 2011 Central Thailand Flood. Remote Sens. 2017, 9, 78. [Google Scholar] [CrossRef]
  28. Qiu, J.; Cao, B.; Park, E.; Yang, X.; Zhang, W.; Tarolli, P. Flood Monitoring in Rural Areas of the Pearl River Basin (China) Using Sentinel-1 SAR. Remote Sens. 2021, 13, 1384. [Google Scholar] [CrossRef]
  29. Martinis, S.; Twele, A.; Voigt, S. Unsupervised Extraction of Flood-Induced Backscatter Changes in SAR Data Using Markov Image Modeling on Irregular Graphs. IEEE Trans. Geosci. Remote Sens. 2011, 49, 251–263. [Google Scholar] [CrossRef]
  30. Matgen, P.; Hostache, R.; Schumann, G.; Pfister, L.; Hoffmann, L.; Savenije, H.H.G. Towards an Automated SAR-Based Flood Monitoring System: Lessons Learned from Two Case Studies. Phys. Chem. Earth 2011, 36, 241–252. [Google Scholar] [CrossRef]
  31. Schlaffer, S.; Matgen, P.; Hollaus, M.; Wagner, W. Flood Detection from Multi-Temporal SAR Data Using Harmonic Analysis and Change Detection. Int. J. Appl. Earth Obs. Geoinf. 2015, 38, 15–24. [Google Scholar] [CrossRef]
  32. Chini, M.; Hostache, R.; Giustarini, L.; Matgen, P. A Hierarchical Split-Based Approach for Parametric Thresholding of SAR Images: Flood Inundation as a Test Case. IEEE Trans. Geosci. Remote Sens. 2017, 55, 6975–6988. [Google Scholar] [CrossRef]
  33. Singha, M.; Dong, J.; Sarmah, S.; You, N.; Zhou, Y.; Zhang, G.; Doughty, R.; Xiao, X. Identifying Floods and Flood-Affected Paddy Rice Fields in Bangladesh Based on Sentinel-1 Imagery and Google Earth Engine. ISPRS J. Photogramm. Remote Sens. 2020, 166, 278–293. [Google Scholar] [CrossRef]
  34. Wu, X.; Zhang, Z.; Zhang, W.; Yi, Y.; Zhang, C.; Xu, Q. A Convolutional Neural Network Based on Grouping Structure for Scene Classification. Remote Sens. 2021, 13, 2457. [Google Scholar] [CrossRef]
  35. Yi, Y.; Zhang, Z.; Zhang, W.; Jia, H.; Zhang, J. Landslide Susceptibility Mapping Using Multiscale Sampling Strategy and Convolutional Neural Network: A Case Study in Jiuzhaigou Region. Catena 2020, 195, 104851. [Google Scholar] [CrossRef]
  36. Jiang, X.; Liang, S.; He, X.; Ziegler, A.D.; Lin, P.; Pan, M.; Wang, D.; Zou, J.; Hao, D.; Mao, G.; et al. Rapid and Large-Scale Mapping of Flood Inundation via Integrating Spaceborne Synthetic Aperture Radar Imagery with Unsupervised Deep Learning. ISPRS J. Photogramm. Remote Sens. 2021, 178, 36–50. [Google Scholar] [CrossRef]
  37. Dong, Z.; Wang, G.; Amankwah, S.O.Y.; Wei, X.; Hu, Y.; Feng, A. Monitoring the Summer Flooding in the Poyang Lake Area of China in 2020 Based on Sentinel-1 Data and Multiple Convolutional Neural Networks. Int. J. Appl. Earth Obs. Geoinf. 2021, 102, 102400. [Google Scholar] [CrossRef]
  38. Bonafilia, D.; Tellman, B.; Anderson, T.; Issenberg, E. Sen1Floods11: A Georeferenced Dataset to Train and Test Deep Learning Flood Algorithms for Sentinel-1. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, 14–19 June 2020; pp. 835–845. [Google Scholar] [CrossRef]
  39. Katiyar, V.; Tamkuan, N.; Nagai, M. Near-Real-Time Flood Mapping Using off-the-Shelf Models with Sar Imagery and Deep Learning. Remote Sens. 2021, 13, 2334. [Google Scholar] [CrossRef]
  40. Konapala, G.; Kumar, S.V.; Khalique Ahmad, S. Exploring Sentinel-1 and Sentinel-2 Diversity for Flood Inundation Mapping Using Deep Learning. ISPRS J. Photogramm. Remote Sens. 2021, 180, 163–173. [Google Scholar] [CrossRef]
  41. Nemni, E.; Bullock, J.; Belabbes, S.; Bromley, L. Fully Convolutional Neural Network for Rapid Flood Segmentation in Synthetic Aperture Radar Imagery. Remote Sens. 2020, 12, 2532. [Google Scholar] [CrossRef]
  42. Ronneberger, O.; Fischer, P.; Brox, T. Computer U-Net: Convolutional Networks for Biomedical Image Segmentation. Int. Conf. Med. Image Comput. Comput. Interv. 2015, 9, 16591–16603. [Google Scholar] [CrossRef]
  43. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef]
  44. Long, J.; Shelhamer, E.; Darrell, T. Fully Convolutional Networks for Semantic Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  45. Badrinarayanan, V.; Kendall, A.; Cipolla, R. SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef]
  46. Chen, L.; Papandreou, G.; Schroff, F.; Adam, H. Rethinking Atrous Convolution for Semantic Image Segmentation. arXiv 2017, arXiv:1706.05587. [Google Scholar] [CrossRef]
  47. Yi, Y.; Zhang, Z.; Zhang, W.; Zhang, C.; Li, W.; Zhao, T. Semantic Segmentation of Urban Buildings from VHR Remote Sensing Imagery Using a Deep Convolutional Neural Network. Remote Sens. 2019, 11, 1774. [Google Scholar] [CrossRef]
  48. Twele, A.; Cao, W.; Plank, S.; Martinis, S. Sentinel-1-Based Flood Mapping: A Fully Automated Processing Chain. Int. J. Remote Sens. 2016, 37, 2990–3004. [Google Scholar] [CrossRef]
  49. Abbasi, M.; Shah-hosseini, R.; Mohammad, A.-N. Sentinel-1 Polarization Comparison for Flood Segmentation Using Deep Learning. Proceedings 2023, 87, 14. [Google Scholar] [CrossRef]
  50. Uddin, K.; Matin, M.A.; Meyer, F.J. Operational Flood Mapping Using Multi-Temporal Sentinel-1 SAR Images: A Case Study from Bangladesh. Remote Sens. 2019, 11, 1581. [Google Scholar] [CrossRef]
  51. Schlaffer, S.; Chini, M.; Giustarini, L.; Matgen, P. Probabilistic Mapping of Flood-Induced Backscatter Changes in SAR Time Series. Int. J. Appl. Earth Obs. Geoinf. 2017, 56, 77–87. [Google Scholar] [CrossRef]
  52. Zhang, L.; Xia, J. Flood Detection Using Multiple Chinese Satellite Datasets during 2020 China Summer Floods. Remote Sens. 2022, 14, 51. [Google Scholar] [CrossRef]
  53. Wang, Y.; Colby, J.D.; Mulcahy, K.A. An Efficient Method for Mapping Flood Extent in a Coastal Floodplain Using Landsat TM and DEM Data. Int. J. Remote Sens. 2002, 23, 3681–3696. [Google Scholar] [CrossRef]
  54. DeVries, B.; Huang, C.; Armston, J.; Huang, W.; Jones, J.W.; Lang, M.W. Rapid and Robust Monitoring of Flood Events Using Sentinel-1 and Landsat Data on the Google Earth Engine. Remote Sens. Environ. 2020, 240, 111664. [Google Scholar] [CrossRef]
Figure 1. Geo-location map of the Yangtze River Basin.
Figure 1. Geo-location map of the Yangtze River Basin.
Remotesensing 17 02909 g001
Figure 2. Workflow for inundation detection and flood mapping in present study. The subfigures (ac) illustrate the detailed experimental procedures.
Figure 2. Workflow for inundation detection and flood mapping in present study. The subfigures (ac) illustrate the detailed experimental procedures.
Remotesensing 17 02909 g002
Figure 3. Examples derived by the semi-automatic method for (a) farmland and aquaculture pond; (b) mountain area; (c) hilly terrain; (d) rivers.
Figure 3. Examples derived by the semi-automatic method for (a) farmland and aquaculture pond; (b) mountain area; (c) hilly terrain; (d) rivers.
Remotesensing 17 02909 g003
Figure 4. Locations of the flood events recorded from 2016 to 2021 in the YRB. Subfigure (a) shows the spatial distribution of all Sentinel data used in the Yangtze River Basin, while subfigures (bf) represent the training and testing datasets, respectively.
Figure 4. Locations of the flood events recorded from 2016 to 2021 in the YRB. Subfigure (a) shows the spatial distribution of all Sentinel data used in the Yangtze River Basin, while subfigures (bf) represent the training and testing datasets, respectively.
Remotesensing 17 02909 g004
Figure 5. Architecture of the FloodsNet. @1, @ 2, @ 3 mean that the dilated rates are 1, 2, and 3, respectively. (For an introduction to dilated rates, please refer to Section Atrous Spatial Pyramid Pooling (ASPP)).
Figure 5. Architecture of the FloodsNet. @1, @ 2, @ 3 mean that the dilated rates are 1, 2, and 3, respectively. (For an introduction to dilated rates, please refer to Section Atrous Spatial Pyramid Pooling (ASPP)).
Remotesensing 17 02909 g005
Figure 6. Illustration of ASPP: (a) Atrous convolution with a dilation rate of 1. (b) Atrous convolution with a dilation rate of 2. (c) Atrous convolution with a dilation rate of 3. (d) The ASPP structure integrated into FloodsNet.
Figure 6. Illustration of ASPP: (a) Atrous convolution with a dilation rate of 1. (b) Atrous convolution with a dilation rate of 2. (c) Atrous convolution with a dilation rate of 3. (d) The ASPP structure integrated into FloodsNet.
Remotesensing 17 02909 g006
Figure 7. Skip connection structure.
Figure 7. Skip connection structure.
Remotesensing 17 02909 g007
Figure 8. Results of ablation experiments. (a) VH polarized band. (b) Label. (c) Baseline. (d) Baseline+Resblock. (e) Baseline+Resblock+ASPP. (f) Baseline+Resblock+SC. (g) FloodsNet.
Figure 8. Results of ablation experiments. (a) VH polarized band. (b) Label. (c) Baseline. (d) Baseline+Resblock. (e) Baseline+Resblock+ASPP. (f) Baseline+Resblock+SC. (g) FloodsNet.
Remotesensing 17 02909 g008
Figure 9. Comparison experiment results of the model. (a) VH polarized band. (b) Label. (c) FCN-8. (d) UNet. (e) SegNet. (f) Deeplabv3. (g) DeepResUNet. (h) FloodsNet.
Figure 9. Comparison experiment results of the model. (a) VH polarized band. (b) Label. (c) FCN-8. (d) UNet. (e) SegNet. (f) Deeplabv3. (g) DeepResUNet. (h) FloodsNet.
Remotesensing 17 02909 g009
Figure 10. Performance comparisons of the model with different combined bands as inputs. (a) VH polarized band. (b) Label. (c) 1. (d) 2. (e) 3. (f) 4. (g) 15. (h) 25. (i) 35. (j) 45. (k) 12. (l) 34. (m) 125. (n) 345. (o) 1234. (p) 12345. 1, 2, 3, 4, and 5 represent five bands of VH_PRE, VV_PRE, VH_ORI, VV_ORI, and DEM, respectively.
Figure 10. Performance comparisons of the model with different combined bands as inputs. (a) VH polarized band. (b) Label. (c) 1. (d) 2. (e) 3. (f) 4. (g) 15. (h) 25. (i) 35. (j) 45. (k) 12. (l) 34. (m) 125. (n) 345. (o) 1234. (p) 12345. 1, 2, 3, 4, and 5 represent five bands of VH_PRE, VV_PRE, VH_ORI, VV_ORI, and DEM, respectively.
Remotesensing 17 02909 g010
Figure 11. Floods detected in the YRB from 2016 to 2021 with our proposed model. Different colors represent flooded areas in different years: (a) Dongting Lake flood from 4 June 2017 to 10 July 2017. (b) The middle reaches of the Yangtze River flood from 11 June 2016 to 15 July 2017. (c) Poyang Lake flood from 20 June to 26 July 2020. (d) The upper reaches of the Yangtze River flood from 16 August 2021 to 21 September 2021. (e) Chaohu Lake flood from 3 July 2020 to 27 July 2020. (f) Huaihe River flood from 7 August 2018 to 19 August 2018.
Figure 11. Floods detected in the YRB from 2016 to 2021 with our proposed model. Different colors represent flooded areas in different years: (a) Dongting Lake flood from 4 June 2017 to 10 July 2017. (b) The middle reaches of the Yangtze River flood from 11 June 2016 to 15 July 2017. (c) Poyang Lake flood from 20 June to 26 July 2020. (d) The upper reaches of the Yangtze River flood from 16 August 2021 to 21 September 2021. (e) Chaohu Lake flood from 3 July 2020 to 27 July 2020. (f) Huaihe River flood from 7 August 2018 to 19 August 2018.
Remotesensing 17 02909 g011
Figure 12. VV and VH polarization.
Figure 12. VV and VH polarization.
Remotesensing 17 02909 g012
Figure 13. Slope statistics of the training and test datasets. (a) Average slope of 5296 training images. (b,c) Probability density figures of slope distribution for 13 images in test dataset1, and for 14 images in test dataset2, respectively. (Note: The slope probability density figures of adjacent images in the test dataset are the same because the images come from the same area before and after the flooding).
Figure 13. Slope statistics of the training and test datasets. (a) Average slope of 5296 training images. (b,c) Probability density figures of slope distribution for 13 images in test dataset1, and for 14 images in test dataset2, respectively. (Note: The slope probability density figures of adjacent images in the test dataset are the same because the images come from the same area before and after the flooding).
Remotesensing 17 02909 g013
Figure 14. Dongting Lake floods in 2017. (a) 30 May. (b) 11 June. (c) 23 June. (d) 5 July. (e) 17 July. (f) 29 July. (g) 10 August. (h) 22 August.
Figure 14. Dongting Lake floods in 2017. (a) 30 May. (b) 11 June. (c) 23 June. (d) 5 July. (e) 17 July. (f) 29 July. (g) 10 August. (h) 22 August.
Remotesensing 17 02909 g014
Figure 15. Poyang Lake floods in 2020. (a) 8 June. (b) 20 June. (c) 2 July. (d) 14 July. (e) 26 July. (f) 7 August. (g) 19 August. (h) 31 August.
Figure 15. Poyang Lake floods in 2020. (a) 8 June. (b) 20 June. (c) 2 July. (d) 14 July. (e) 26 July. (f) 7 August. (g) 19 August. (h) 31 August.
Remotesensing 17 02909 g015
Figure 16. The variation of flooded areas with time on Poyang and Dongting Lakes.
Figure 16. The variation of flooded areas with time on Poyang and Dongting Lakes.
Remotesensing 17 02909 g016
Table 1. Sentinel-1 SAR images available for flooding events from 2016 to 2021 in the YRB. Each image used for training and testing purposes is appropriately labeled. ‘Application’ denotes the utilization of the trained model for direct image prediction without labeled data.
Table 1. Sentinel-1 SAR images available for flooding events from 2016 to 2021 in the YRB. Each image used for training and testing purposes is appropriately labeled. ‘Application’ denotes the utilization of the trained model for direct image prediction without labeled data.
Flood EventsDistrictProduct IDDateTrain or Test
1Dongting Lake011CB0_5A052016/06/09 Train and Test
0127C5_F17D2016/07/03Train and Test
2Poyang Lake011822_A9282016/05/30 Train and Test
012E7C_86E82016/07/17Train and Test
3Middle reaches of the YRB011D9A_08012016/06/11Train and Test
0128B9_886D2016/07/05 Train and Test
4Poyang Lake00A8F1_76322017/06/12 Train and Test
00B2FB_30912017/07/06Train and Test
5Juzhang River02747C_139D2018/07/05Train and Test
027F52_9A162018/07/29 Train and Test
6Huaihe River02836E_FDCB2018/08/07 Train and Test
028919_2CEB2018/08/19Train and Test
7Middle reaches of the YRB032778_40DE2019/07/02 Train and Test
032CC4_6DEF2019/07/14Train and Test
8Ruan Jiang03E75D_6DAE2020/07/30Test
03ED1D_5ADE2020/08/11 Test
9Dongting Lake01C150_503B2017/06/04 Application
01D14B_23A92017/07/10Application
10Poyang Lake029F8B_298B2020/06/20 Application
02AF8A_BF3A2020/07/26Application
11Chaohu Lake03DB5D_91FD2020/07/03Application
03E612_6A3E2020/07/27 Application
12Fujiang River03EE9F_8BAF2020/08/14Application
04012C_41B22020/09/19 Application
13Dongting Lake03D52B_49F42020/06/19 Application
03E52E_6E8E2020/07/25Application
14Middle and lower reaches of the YRB03D2E0_90D32020/06/14 Application
03DD85_0B972020/07/08 Application
15Middle and lower reaches of the YRB03D2E0_261F2020/06/14 Application
03DD85_725A2020/07/08Application
16Upper reaches of the YRB04A272_97F52021/08/16Application
04B46E_4D612021/09/21Application
Table 2. The Comparison Models.
Table 2. The Comparison Models.
ModelsReferencesCharacteristic
FCN-8[44]The first CNN segmentation model uses deconvolution instead of fully connected layers.
UNet[42]Symmetric encoder decoder architecture and skip connection design.
SegNet[45]Its decoder’s use of pooling indices for upsampling, enabling precise segmentation while maintaining low computational overhead and model size.
DeepLab-v3[46]Its Atrous Spatial Pyramid Pooling (ASPP) module captures multi-scale context
DeepResUNet[47]Reducing model parameters while ensuring segmentation accuracy
Table 3. Confusion matrix.
Table 3. Confusion matrix.
Confusion MatrixLabel
PositiveNegative
PredictPositiveTPFP
NegativeFNTN
Table 4. Experimental parameter settings.
Table 4. Experimental parameter settings.
ParametersSetup
OptimizerAdam
Batch size10
Training times60,000
Initial learning rate0.0001
Decay strategyExponential decay
Decay frequency10,000 times/0.8
Table 5. Results of ablation experiments conducted with two distinct test datasets. The first line of each model represents the evaluation metrics obtained with test dataset 1, while the second line represents those with test dataset 2. Values highlighted in bold indicate the highest numbers for corresponding metrics obtained.
Table 5. Results of ablation experiments conducted with two distinct test datasets. The first line of each model represents the evaluation metrics obtained with test dataset 1, while the second line represents those with test dataset 2. Values highlighted in bold indicate the highest numbers for corresponding metrics obtained.
ModelOAPrecisionRecallF1_scoreKappa
Baseline0.9800.9930.9380.9650.951
0.9700.9860.8670.9230.904
Baseline +
Resblock
0.9820.9910.9480.9690.956
0.9750.9840.8930.9370.921
Baseline + Resblock + ASPP0.9870.9840.9730.9780.969
0.9810.9680.9380.9530.941
Baseline + Resblock + SC0.9850.9850.9650.9750.965
0.9790.9700.9260.9470.934
FloodsNet0.9900.9940.9720.9830.976
0.9850.9870.9400.9630.954
Table 6. Comparison results of model experiments with two test datasets. The first line displays the results of each model with test dataset 1, while the second line corresponds to dataset 2. Values highlighted in bold indicate the highest numbers for the respective metrics.
Table 6. Comparison results of model experiments with two test datasets. The first line displays the results of each model with test dataset 1, while the second line corresponds to dataset 2. Values highlighted in bold indicate the highest numbers for the respective metrics.
ModelOAPrecisionRecallF1_scoreKappa
FCN-80.9740.9430.9700.9560.937
0.9610.8810.9390.9090.884
UNet0.9860.9800.9730.9760.966
0.9780.9510.9420.9470.933
SegNet0.9830.9910.9530.9710.960
0.9750.9810.8970.9370.922
DeepLabv30.9850.9880.9610.9740.964
0.9790.9760.9180.9460.933
DeepResUNet0.9860.9850.9670.9760.966
0.9790.9700.9270.9480.935
FloodsNet0.9900.9940.9720.9830.976
0.9850.9870.9400.9630.954
Table 7. Performance comparisons across various band combinations using two specific test datasets. The bands are denoted as 1, 2, 3, 4, and 5, representing VH_PRE, VV_PRE, VH_ORI, VV_ORI, and DEM, respectively. The first line displays the performance metrics for each model when provided with band combinations from test dataset 1, while the second line corresponds to test dataset 2. Values in bold highlight the highest scores for the respective metrics.
Table 7. Performance comparisons across various band combinations using two specific test datasets. The bands are denoted as 1, 2, 3, 4, and 5, representing VH_PRE, VV_PRE, VH_ORI, VV_ORI, and DEM, respectively. The first line displays the performance metrics for each model when provided with band combinations from test dataset 1, while the second line corresponds to test dataset 2. Values in bold highlight the highest scores for the respective metrics.
BandOAPrecisionRecallF1_scoreKappa
10.9530.9500.8900.9190.886
0.9780.9560.9090.9320.915
20.9540.9330.9090.9210.888
0.9560.9070.8790.8920.865
30.9900.9940.9720.9830.976
0.9850.9870.9400.9630.954
40.9760.9670.9510.9590.942
0.9630.9470.8710.9070.885
1, 50.9550.9430.9030.9230.891
0.9720.9210.9450.9330.915
2, 50.9520.9320.9030.9170.883
0.9550.9010.8800.8900.862
3, 50.9860.9870.9630.9750.965
0.9790.9730.9230.9470.934
4, 50.9750.9620.9550.9580.941
0.9620.9370.8720.9030.879
1, 20.9580.9360.9210.9280.898
0.9630.9120.9100.9110.888
3, 40.9820.9710.9630.9690.957
0.9700.9630.8860.9230.904
1, 2, 50.9460.9460.8690.9050.868
0.9610.9120.8990.9060.881
3, 4, 50.9810.9690.9670.9680.954
0.9700.9510.8990.9240.905
1, 2, 3, 40.9790.9700.9600.9650.950
0.9650.9130.9160.9150.893
1, 2, 3, 4, 50.9800.9690.9620.9650.951
0.9650.9110.9200.9160.894
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wu, X.; Zhang, Z.; Zhang, W.; An, B.; Li, Z.; Li, R.; Chen, Q. Large-Scale Flood Detection and Mapping in the Yangtze River Basin (2016–2021) Using Convolutional Neural Networks with Sentinel-1 SAR Images. Remote Sens. 2025, 17, 2909. https://doi.org/10.3390/rs17162909

AMA Style

Wu X, Zhang Z, Zhang W, An B, Li Z, Li R, Chen Q. Large-Scale Flood Detection and Mapping in the Yangtze River Basin (2016–2021) Using Convolutional Neural Networks with Sentinel-1 SAR Images. Remote Sensing. 2025; 17(16):2909. https://doi.org/10.3390/rs17162909

Chicago/Turabian Style

Wu, Xuan, Zhijie Zhang, Wanchang Zhang, Bangsheng An, Zhenghao Li, Rui Li, and Qunli Chen. 2025. "Large-Scale Flood Detection and Mapping in the Yangtze River Basin (2016–2021) Using Convolutional Neural Networks with Sentinel-1 SAR Images" Remote Sensing 17, no. 16: 2909. https://doi.org/10.3390/rs17162909

APA Style

Wu, X., Zhang, Z., Zhang, W., An, B., Li, Z., Li, R., & Chen, Q. (2025). Large-Scale Flood Detection and Mapping in the Yangtze River Basin (2016–2021) Using Convolutional Neural Networks with Sentinel-1 SAR Images. Remote Sensing, 17(16), 2909. https://doi.org/10.3390/rs17162909

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop