1. Introduction
Natural disasters, such as floods, have caused extensive damage since they occur frequently, while climate change exacerbates their effects, thereby endangering human lives and both infrastructure and ecosystems worldwide. The past five years have witnessed major flood disasters in Spain (2024), Pakistan (2022), Türkiye (2021), Western Europe (2021), and China (2021), which have caused extensive economic damage and human loss and humanitarian disasters worldwide. The 2021 Limburg floods demonstrated the critical need for better flood monitoring and risk mitigation approaches when they led to almost €600 million in damages throughout the Netherlands, Belgium, and Germany [
1,
2]. The Netherlands faces extreme flooding risks because its territory extends below sea level by a quarter while maintaining high population density and infrastructure development levels in Europe [
3]. The country experiences severe economic and social impacts from flooding despite having globally recognized flood control systems. Therefore, the need to develop precise and timely flood extent information becomes essential, since it supports both emergency response operations and long-term climate adaptation planning.
Remote sensing (RS) technologies provide innovative solutions to the constraints of ground-based monitoring systems because they allow for wide-area data acquisition at repeatable intervals and in near real time. The technology of RS functions as an efficient and adaptable instrument for flood surveillance, as it provides broad spatial coverage while enabling frequent temporal monitoring. RS platforms utilize satellite-based and aircraft-based systems and Unmanned Aerial Vehicles (UAVs) to deliver key data required to create and monitor flood extents. The two fundamental technologies used in these platforms are Synthetic Aperture Radar (SAR) and optical sensors. SAR sensors remain essential for flood detection because they can penetrate through cloud cover and obtain images at any time of day while maintaining reliability under various weather conditions [
4,
5,
6]. Multiple SAR sources like Sentinel-1, the Advanced Land Observing Satellite (ALOS), and Radarsat can enhance temporal resolution in flood surveillance through integration. On the other hand, commercial SAR data can be expensive; however, the industry needs effective ways to integrate various SAR sensors which differ in configuration. As their drawbacks, SAR-based flood detection experiences difficulties in separating water from dark surfaces such as snow and ice, as well as deserts and dry croplands, because these surfaces produce similar backscatter intensities [
7]. In contrast, the optical imaging system benefits from strong land cover discrimination in the visible and infrared regions as well as high signal-to-noise ratios and multiple spectral bands, which makes it suitable for flood and water body monitoring [
8]. Nevertheless, cloud cover remains a major limitation for optical sensors due to obstructing the visibility during crucial flood events. Moreover, water reflectance might show considerable variations because of turbidity along with vegetation presence and sun angle conditions, which makes accurate flooded area delineation difficult, especially in high-resolution images [
7,
9]. Given these challenges, the use of single-sensor approaches with either SAR or optical systems can produce incomplete or unreliable flood detection outcomes when dealing with complex and dynamic landscapes.
To overcome the limitations of single-sensor methods, the fusion or combined use of SAR and optical data has become an increasingly effective approach in flood mapping [
10,
11]. As highlighted above, SAR provides robust geometric and surface roughness information in all weather conditions alongside optical imagery, which enhances land–water discrimination by using spectral indices. The Normalized Difference Water Index (NDWI) and Modified NDWI (MNDWI) operate as leading optical indices for detecting open water bodies across both natural and urban areas by using different band combinations of Green-Near Infrared (NIR) and Green-Shortwave Infrared (SWIR) [
12,
13]. Flood-affected vegetation can be detected based on the Normalized Difference Vegetation Index (NDVI) because it reveals plant health and canopy greenness anomalies [
14]. The Normalized Difference Moisture Index (NDMI) and Land Surface Water Index (LSWI) provide improved sensitivity to both surface moisture and shallow water in agricultural areas, according to Stoyanova [
15] and Mohammadi et al. [
16]. The Hue-Saturation-Value (HSV) color space transformation enables better separation of water from spectrally similar shadows or dark rooftops in urban areas; also, the HSV provides effective solutions for flood monitoring [
17]. Compared to the traditional water indices (i.e., NDWI, AWEI), HSV’s better performance is likely due to its more effective spectral alignment with water-related features. Ref. [
18] recommended the HSV color space instead of Red-Green-Blue (RGB) due to reasons such as the existence of correlation between bands in the RGB space and the difficulty of selecting a specific color in threshold-based approaches. In the study conducted for the detection of shadow areas in high-resolution images, the HSV space achieved better results than the RGB [
19]. These optical spectral indices enable a detailed analysis of flood extent throughout different land cover types under various illumination situations.
In addition to these optical spectral data, SAR systems provide essential backscatter and structural data to improve flood detection when cloudy or nighttime conditions prevent optical observations. Floodwaters in radar imagery show minimal backscattering because of their reflective properties, leading to the use of coefficients from different polarizations. SAR polarization variations enable the detection of flooded vegetation and the separation of flooded vegetation from other low-backscatter features. Additionally, temporal coherence between SAR acquisitions can reveal abrupt surface changes linked to inundation, while texture metrics derived from the Gray-Level Co-occurrence Matrix (GLCM)—such as contrast, entropy, and homogeneity—capture spatial heterogeneity, which is useful for mapping floodwaters amidst buildings or vegetation. The GLCM texture information, as defined by [
20], describes the co-occurrence of pixel information with different spatial relationships. Several texture information can be generated with different window sizes and angle values. But it can increase the number of input data. On the other hand, the dimension of the data size can be reduced, and a smaller number of features can be selected based on principal component analysis (PCA). The impact on the classification can be considerable, even if the textural information is sparse [
21]. Ref. [
22] proposed dimension reduction using PCA to reduce the number of features produced by GLCM, and they noticed the improvement in the accuracies. Overall, the combination of SAR and optical features creates a multi-dimensional view of flooded areas, which enhances detection accuracy and reliability in complex environments beyond what one sensor can achieve.
The translation of the RS observations into actionable flood hazard information requires various modeling techniques due to the complex nature of flood dynamics and the multiple affected landscapes. The flood hazard mapping methodologies include physical modelling, physically based modelling, and empirical modelling, according to Mudashiru et al. [
23] and Teng et al. [
24]. Physical modeling conducts laboratory tests that duplicate flood processes to study hydrodynamic behaviors, yet these approaches face limitations from scale constraints and high costs [
25,
26]. On the other hand, physically based models apply mathematical expressions for hydrologic and hydraulic processes such as rainfall-runoff transformation, channel flow, and surface inundation that integrate high-resolution spatial data including Digital Elevation Models (DEMs), land use maps, and meteorological inputs [
27]. These models present more realistic representations, yet need complex calibration, consume significant computational resources, and face data availability challenges across various regions worldwide [
26]. In contrast, empirical models utilize statistical or machine learning techniques to build connections between observed flood areas and predictor variables that originate from remote sensing data as well as topography, soil type, and land cover [
24]. The empirical approaches have gained attention in recent years because they base their assessments on data-driven approaches, which provide fast and flexible hazard evaluations without requiring detailed hydrological inputs or boundary conditions. In particular, empirical models utilizing Random Forest (RF), Support Vector Machines (SVMs), and Deep Learning (DL) architectures achieve robust classification when combined with multi-sensor RS data to perform well across different geographic and climatic regions [
28,
29,
30,
31]. Ultimately, the choice of modeling approach depends on the specific objectives, data availability, and computational resources of a given study.
This proposed research aims to develop an integrated framework for flood hazard mapping by using multi-temporal SAR and optical RS data with an RF classifier in heterogeneous landscapes. The research uses freely available Sentinel-1 and Landsat-8 imagery as well as commercial Radarsat-2 data to extract a wide range of flood-sensitive features in the Netherlands, which faces high flood risks despite its well-developed water management systems. The inventory includes SAR backscattering coefficients (, , ), Landsat-derived NDWI, conversion of RGB images to the color space (HSV), and textural components from the GLCM (textPC1). We created 25 different scenarios that span from using single-source data to complex multi-sensor configurations that combine raw and textural information from all three satellite platforms. This study has two main objectives: (1) Advancing the scientific foundation for selecting sensor combinations and features best suited for flood detection in delta regions and (2) Creating functional and extendable operational flood monitoring system approaches. Additionally, we would like to address the following questions: (1) Which feature combinations from Sentinel-1, Radarsat-2, and Landsat-8 (e.g., backscattering coefficients, spectral indices, and texture features) yield the highest accuracy in flood extent mapping? (2) What is the contribution of commercial Radarsat-2 data to flood mapping performance when integrated with Sentinel-1 and Landsat-8? (3) How does the SAR-derived texture metric (GLCM-based textPC1) improve the delineation of flooded areas in heterogeneous and complex landscapes? (4) How effective is the RF classifier in handling multi-source input features and maintaining robustness across diverse flood scenarios and terrain conditions?
3. Results
The classified maps presented in
Figure 3 provide a visual comparison of how different feature combinations affect flood mapping performance for the Zutphen (Flooded Area-A) region in all 25 scenarios. The first three single-feature scenarios (S1–S3) produced noisy results with inconsistent classification of flooded areas throughout urban and cropland zones using Sentinel-1 or Radarsat-2 backscatter data. The Radarsat-2 HH backscatter-based S3 scenario produced the worst results, which resulted in a complete loss of flood detection capabilities. S4–S5 optical-only scenarios showed better spatial patterns with NDWI (S4) improving flood extent detection but also confusing flooded cropland with water bodies and HSV (S5) better capturing urban areas but poorly representing flooding near infrastructure. The addition of SAR textures in S6–S13 led to enhanced spatial coherence together with lower classification noise levels. Notably, S10 produced large negative outcomes because Radarsat-2 HH texture features fail to separate land cover types correctly in complex environments, particularly in croplands and vegetated zones. In contrast, S12 presented the most refined and structured flood map with precise riverbank flood boundary alignment, since it uses all SAR backscatter and texture features. The addition of optical indices in S14–S20 further improved map quality. In particular, the combination of SAR and optical data in S17, S18, and S20 scenarios achieved accurate and balanced mapping results. Among Radarsat-2 and optical combinations (S21–S23), S23 represented the most stable classification outcome between Radarsat-2 and optical combinations because it achieved clear flood delineation. Finally, S25, which integrated all SAR and optical features along with texture metrics, delivered the highest visual quality. The visual performance matched the quantitative excellence because full multi-sensor fusion with texture information produced the most reliable flood mapping results in complex urban–rural landscapes.
The 25 scenarios applied to the Heijen region (Flooded Area-B) are presented in
Figure 4 to visually evaluate classification quality in a primarily agricultural landscape. The three single-feature scenarios (S1–S3) showed major misclassification problems. Sentinel-1 backscatter inputs (S1, S2) generated maps with numerous incorrect flood detections throughout cropland and vegetated areas, and Radarsat-2 HH backscatter (S3) fails again to detect flood zones by showing excessive homogeneity of cropland. The optical-only scenarios produced better visual results. The NDWI (S4) yielded accurate flood boundary detection near river channels but incorrectly labeled some agricultural land as water bodies. The HSV (S5) method provided better identification of urban and tree-covered regions yet failed to detect flooding, particularly in vegetated areas. The visual improvement from SAR texture-based scenarios (S6–S13) remained less effective than what was observed in Zutphen. The poor performance of S10 continued from Area A by incorrectly expanding the urban class while hiding flooded cropland. On the other hand, the addition of texture in S11 and S12 led to better detection of flooded zones, although some noise remained present. The introduction of optical data in scenarios S14–S20 contributed to much better flood delineation. For instance, the maps produced by S18 and S20 show balanced visual results because they correctly depicted hydrological patterns and minimized misclassifications. Among Radarsat-2 and optical combinations (S21–S23), S23 produced the best results by precisely mapping flood areas and differentiating between trees. In the end, the full-feature scenario S25 delivered the most accurate visual classification results. The visual coherence and flood delineation quality of S25 supported its leading performance metrics, which demonstrated that multi-sensor texture-augmented approaches worked equally well in agricultural landscapes.
Considering the general classification accuracy assessment, the effect of band combinations obtained with different sensors and indices on the accuracy was analysed in two areas, and their accuracy was evaluated with OA and Kappa (
Figure 5,
Table 4). In the single-band scenarios (S1–S5), S5 (Landsat HSV) gave the best result with an OA of 86.6% and 80.7% in Zutphen and Heijen, respectively. Among the combinations with Radar data only (between S6 and S13), the best scenario was obtained with S12 for both areas (
Figure 5,
Table 4). In S14, where Landsat HSV and Landsat NDWI were used together, the OA in Flooded Area-A increased from 67.6% to 92.7% compared to the NDWI-only scenario (S4). On the other hand, in Flooded Area-B for S14, OA increased from 77.1% to 81.2% compared to the NDWI-only scenario (S4). For Sentinel-1 and optical combinations (S15–S20), OA in Flooded Area-A ranged from 86.2% to 96.4%, and in Flooded Area-B, OA ranged between 80.2% and 86.7%. In the scenarios where Radarsat-2 and Landsat-8 were used together (for S21–S23), the highest accuracy was S23 with OA 93.3% for Flooded Area-A. On the contrary, for Flooded Area-B, the highest OA was S21 with 83.1%. In Flooded Area-A, the use of Landsat-8 HSV increased the accuracy by about 2%. In scenarios S24 and S25, which include the combination of all data, S25 provided higher accuracies. The presence of texture information in S25 improved the overall accuracy in Zupthen and Heijen by 2% and 6%, respectively. S25 Landsat HSV was better than Landsat NDWI for both areas (
Figure 5,
Table 4). The best band combinations (S20 and S25) provided 96.4–96.3% OA and 0.950–0.949 Kappa for Flooded Area-A (Zutphen), respectively. For the Flooded Area-B (Heijen), S25 presented the best results with 87.5% OA and 0.833 Kappa.
F1 scores were calculated for class accuracies in both areas (
Figure 6,
Figure 7 and
Table 5,
Table 6). In Area-A (Zutphen), in all scenarios, the Flooded area class was higher than 0.742, the Urban class was higher than 0.529, the Cropland class was higher than 0.651, and the Trees class of S5 and S14–S25 was higher than 0.85 (
Figure 6,
Table 5). In the evaluations conducted using a single feature (S1–S5), optical data provided satisfactory accuracies in both test sites for the flooded area class. This is provided by HSV for Area-A and NDWI for Area-B.
When data combinations are compared, both the SAR data among themselves and the SAR and optical combinations increased the accuracy on a class basis. For Area A, the highest F1 values for flooded areas and trees were obtained with S25 as 0.969 and 0.971, respectively. For urban and cropland, F1 values were found as 0.969 and 0.968 in the S20 combination, respectively. For these classes, the best values after S20 were again obtained with S25. Only the water class was lower than the others, yet, with the F1 score of 0.778, the S23 combination provided the highest accuracy in this class.
In Area-2, only SAR features (S11 and S12) provided the best results (both F1 = 0.969) for the flooded area even though the overall classification accuracies of these scenarios remained lower (
Figure 7,
Table 6). However, NDWI (S4) also provided an F1 score of 0.949, which was very close to the best one. Concerning the Urban class, HSV of Landsat (S5) provided high accuracy. In the Trees class, S5, S14, S16, S17, S19, S20, and S22–S25 provide higher F1 values than 0.990. For the cropland class, different scenarios, such as S14, S16, S17, S19, S20 and S22–S25, respectively, presented high accuracies. Concerning the water class that represents permanent water boundaries, few scenarios like S4, S15, and S21 achieved the F1 score above 0.330.
To visually assess the performance of flood classification under different multi-source data configurations, comparisons were made between the classification outputs of Scenarios 2, 3, 4, and 25 and high-resolution aerial orthophotos obtained from
https://geotiles.citg.tudelft.nl/: accessed on 1 January 2025 (
Figure 8). This analysis includes detailed insets of flooded areas to demonstrate how multi-source data fusion enhances the detection of riverbanks, small ponds, and floodplain boundaries. The results demonstrate that boundary accuracy and small water body detection improve with an increase in the number and variety of input factors. The floodwater extents in Scenario 25 achieved high spatial fidelity since it combined optical and SAR data with terrain features to distinguish between permanent water bodies and temporary flooded zones. The reduced number of mixed pixels at class boundaries demonstrated how data fusion improved classification sharpness. Eventually, the multi-source approaches demonstrate their ability to detect hydrological features that single-source methods usually overlook.
4. Discussion
This research evaluates the multi-sensor flood mapping approach through Sentinel-1, Radarsat-2, and Landsat-8 data integration with the RF classification in two different Dutch landscapes. The study reveals that the best feature combinations differ between the urban Zutphen (Flooded Area-A) and the agricultural Heijen (Flooded Area-B) test sites, with a specific focus on flood class detection through F1 scores and complete classification performance evaluation using the OA and Kappa coefficients.
4.1. The Best Feature Combinations from Sentinel-1, Radarsat-2, and Landsat-8 for the Highest Accuracy in Flood Extent Mapping
The multi-sensor fusion approach (S25) proved to be the most balanced solution because it performed best in both flood detection (F1 = 0.969 in Zutphen, 0.859 in Heijen) and general classification (96.3% OA and 0.949 Kappa in Zutphen; 87.5% OA and 0.833 Kappa in Heijen). Some scenarios demonstrated significant differences between their overall and flood-specific performance results. For example, S16 (Sentinel-1 VV/VH + Landsat HSV) achieved excellent overall metrics (93.3% OA and 0.909 Kappa) in Zutphen but had a relatively lower flood F1 score (0.918), which indicates that the general classification was strong but some flood pixels were misclassified as urban features because of spectral similarities. The classification accuracy of S14 (Landsat NDWI + HSV) in Heijen was 81.2% OA and 0.756 Kappa, but it had the poorest flood F1 score (0.649) among multi-feature scenarios, which shows that the spectral indices performed well for general land cover classification, yet failed to detect floods in agricultural areas, possibly because of mixed pixel effects in croplands.
The urban environment of Zutphen achieved its best results through the combination of SAR texture features with Landsat’s HSV transformation (S19: F1 = 0.967, OA = 96.1%, Kappa = 0.946), where HSV proved its efficiency for differentiating floodwaters from urban features. The tradeoff became apparent in S5 (Landsat HSV alone) because it achieved 86.6% OA and 0.818 Kappa in Zutphen; however, its flood F1 score reached only 0.811. This shows that optical data can work well for general classification but needs SAR enhancement to reach its best performance in flood mapping. The agricultural area of Heijen showed multiple scenarios that achieved high results for both flood identification and total classification accuracy. S15 (Sentinel-1 VV/VH + Landsat NDWI) achieved outstanding performance through its excellent flood detection (F1 = 0.947) and strong overall metrics (85.4% OA/0.798 Kappa), and S18 (Sentinel-1 VV/VH + textures + Landsat NDWI) achieved even better results (F1 = 0.967, 86.7% OA/0.814 Kappa). The flood detection capability of Radarsat-2’s HH polarization with NDWI (Scenario S21: F1 = 0.951, OA = 83.1%, Kappa = 0.767) matches S15 and S18, but its overall classification performance is slightly lower. The high F1 score of Radarsat-2’s HH polarization for agricultural flood detection indicates its effectiveness; nevertheless, it provides limited discrimination between different land cover classes [
50]. The results indicate that the best sensor combinations depend on whether the goal is to achieve maximum flood detection precision or to maintain fair performance across all land cover types.
The combination of NDWI with SAR backscatter and textures produced the best results because NDWI improved water detection while SAR backscatter and textures preserved good discrimination of other land cover classes [
12,
51]. The NDWI scenario (S4) in Heijen produced an opposite pattern with 77.1% OA and a 0.691 Kappa value but achieved an outstanding flood F1 score of 0.949, which indicates that NDWI performs well for flood feature detection but struggles with other land cover type identification in complex agricultural areas [
13,
52]. The differences between these results demonstrate why it is crucial to choose features depending on the main goals of the application between complete land cover classification and exact flood boundary mapping.
4.2. The Contribution of Commercial Radarsat-2 Data to Flood Mapping Performance
The analysis of Radarsat-2’s contribution revealed important but limited benefits that do not universally justify its mandatory inclusion. The data showed site-specific advantages; however, the overall impact was not substantial enough to recommend its definitive use across all scenarios. It is also crucial to highlight that the implications of Radarsat-2 in this study are based on the resampled pixel (30 m) values, not on the native resolution. The urban Zutphen area showed only small improvements in flood detection through Radarsat-2’s HH polarization (S7: F1 = 0.856 vs. S6: F1 = 0.847) with no significant impact on overall classification metrics (77.0% OA/0.687 Kappa vs. 73.6% OA/0.642 Kappa, respectively). On the other hand, the agricultural environment of Heijen showed more noticeable benefits, especially for flooded vegetation detection (S7: F1 = 0.955 vs. S6: F1 = 0.928), although these came with only modest gains in overall accuracy (68.2% vs. 61.7% OA, respectively).
The evaluation of these enhancements requires consideration of three essential factors: (1) the commercial price and restricted time availability of Radarsat-2 data compared to Sentinel-1 data, which is freely available; (2) the small performance gains that could be obtained through different feature combinations without Radarsat-2 data (e.g., S18 in Heijen reached F1 = 0.967 without Radarsat-2); and (3) the limited effect on general land cover classification when focusing on flood detection tasks. This research indicates that Radarsat-2 delivers additional value in particular agricultural flood mapping applications, yet its use is not sufficient to achieve high-quality results. Operational systems could select Sentinel-1 and optical data combinations as their primary choice while using Radarsat-2 only for specific cases where its HH polarization can solve particular ambiguities in vegetated flood detection [
53].
4.3. The Contribution of SAR-Derived Texture Metric (GLCM-Based textPC1) for the Delineation of Flooded Areas
The SAR-derived texture metric (GLCM-based textPC1) enhanced flood delineation in both urban and agricultural landscapes when combined with various SAR backscatter features, but the degree of enhancement depended on polarization and land cover type. The addition of texture features to Sentinel-1 VV backscatter (S8) improved flood detection in urban Zutphen to F1 = 0.916 compared to VV backscatter alone (S1: F1 = 0.798), which resulted in a 14.8% increase and improved overall accuracy from 69.6% to 77.7%. The addition of textures to VH polarization (S9) showed a similar improvement but less than VV polarization (S9: F1 = 0.873 vs. S2: F1 = 0.806), indicating VV’s superior urban flood discrimination ability [
54]. The agricultural landscape of Heijen demonstrated the most significant texture benefits for Radarsat-2’s HH polarization (S10: F1 = 0.955 vs. S3: F1 = 0.940), which resulted in a slight improvement but substantially reduced overall accuracy (46.4% vs. 57.3%). The results show that texture features enhance flood detection ability across all SAR bands and polarizations, but their effectiveness depends on sensor settings and environmental conditions [
55]. The flood-specific metrics (F1 scores) showed consistent improvement despite variable overall classification accuracy, since texture features excel at water detection rather than land cover separation, which makes them essential for operational flood mapping [
56].
4.4. The Effectiveness of the RF Classifier in Handling Multi-Source Input Features and Maintaining Robustness Across Diverse Flood Scenarios and Terrain Conditions
The RF classifier achieved strong results for flood detection at both test sites because its F1 scores for the flooded class exceeded 0.90 in most multi-sensor scenarios. The classifier not only achieved outstanding flood detection performance (F1 = 0.969 in S25) but also maintained high overall classification accuracy (96.3% OA) in Zutphen. Similarly, the classifier yielded excellent flood detection (best F1 = 0.969 in S11 and S12) in Heijen even though the overall classification metrics reached 76.6%, which indicates the classifier focused on precise flood detection rather than the other class accuracies.
4.5. Hyperparameter Optimization: Importance of the Tuning Process
Our RF classifier underwent complete hyperparameter optimization, which produced vital information about model performance across different sensor combinations and landscape types. The evaluation of 36 parameter configurations through testing bagging fractions (0.5, 0.7, 1.0), minimum leaf samples (1, 5, 10), and tree counts (50, 100, 150, 200) showed substantial variations in optimal settings between different scenarios and test sites (
Table 7). The model behavior depends on the intricate connection between input features and landscape characteristics, which results in these observed differences.
Our analysis revealed multiple distinct patterns in the data. The most successful configurations in urban Zutphen used moderate parameter values where minLeafs = 5 appeared in 40% of top scenarios and trees = 150 appeared in 35% of cases (
Table 8). The results indicate that urban flood mapping requires slightly limited model complexity to process diverse features without excessive overfitting. The agricultural Heijen area demonstrated a preference for aggressive splitting through minLeafs = 1 in 45% of scenarios and smaller ensemble sizes with trees = 50 in 40% of cases because simpler models work well for rural flooding patterns. The best-performing scenarios in both areas selected similar model parameters with bag = 0.7–1.0 and minLeafs = 1–5 and trees = 150–200, although agricultural areas reached good performance with fewer trees, which indicates that less computationally intensive models could work for simpler landscapes.
The analysis of sensor inputs in relation to optimal parameters delivered significant findings. The use of optical indices (NDWI/HSV) in scenarios needed more cautious parameter settings for urban areas because these features share similar spectral characteristics. The best results in agricultural areas came from SAR-only scenarios that used texture features when they applied aggressive splitting techniques to detect backscatter variations in flooded vegetation. The performance of texture-enhanced scenarios depended on landscape type because urban applications required 100-tree depth with leaf size set to 5, while agricultural applications needed 200-tree depth with leaf size limited to 1. The optically heavy scenarios showed a consistent preference for minLeafs = 5 in urban areas, yet displayed greater variability in agricultural settings.
The research findings provide essential operational value for flood mapping systems. The results indicate that operational parameter strategies need adaptation because urban applications require constrained growth and larger ensembles; however, agricultural area monitoring can need simpler models. The large differences in optimal parameters between scenarios demonstrate that sensor combination changes require individual parameter adjustments. The study shows that RF optimization depends strongly on both the input data features and the characteristics of the target landscape. The results confirm the necessity of systematic hyperparameter testing for multi-sensor flood mapping systems and indicate potential development of automated optimization methods that adjust parameters through real-time feature and landscape analysis. Future research should investigate adaptive methods to optimize operational implementation without compromising classification accuracy.
5. Conclusions
This study revealed that using Sentinel-1 with Radarsat-2 and Landsat-8 multi-sensor satellite data through RF classification produces accurate and reliable flood extent maps in diverse environments. We evaluated 25 different feature scenarios to analyze the impact of radar backscattering coefficients together with optical indices (NDWI and HSV) and SAR-derived texture metrics (GLCM-based textPC1) on flood extent detection. The study examined two different flood-prone areas in the Netherlands through Zutphen’s urban district and Heijen’s agricultural region.
The findings highlighted that the S25, which combined all SAR and optical features with texture components, produced the highest classification accuracy and flood-specific detection outcomes. The S20 and S25 approaches achieved the highest OAs (96.4–96.3%) and Kappa coefficients (0.950–0.949) along with the highest F1 scores (0.968–0.969) for flooded areas within the complex urban-rural interfaces of Zutphen. On the other hand, the overall classification accuracy of Heijen remained lower at 87.5% with a Kappa coefficient of 0.833 for S25 compared to the Zutphen area but achieved the best performance among the others. Maximum F1 scores of 0.969 for flooded regions in Heijen were obtained through the addition of texture features in scenarios S11 and S12. The research demonstrated an important observation that high-precision flood detection becomes possible through spectral ambiguity reduction in agricultural areas even when land cover classification remains challenging.
SAR texture features proved essential for improving flood delineation within both study areas. The implementation of texture metrics led to improved classification outputs by better differentiating between flooded vegetation and built-up surfaces that resembled spectrally identical background classes. Commercial Radarsat-2 imagery delivered inconsistent performance benefits to the study. The detection of flooded croplands received some improvement through Radarsat-2 HH polarimetric data; however, the system’s total accuracy gain remained relatively minor compared to Sentinel-1 and Landsat-8 freely accessible data. Radarsat-2 should only be used as an optional choice because its operational costs and data availability need to be considered. In addition, other polarizations of the satellite should be tested in future studies.
The RF performed well across different scenarios and landscapes because of its strong adaptability and comprehensive hyperparameter optimization process. The multi-source input processing capability of the classifier led to successful classification results, while specific adjustments to model parameters for each landscape proved essential for achieving optimal detection accuracy. Landscapes with diverse characteristics need adaptive modeling approaches that address both sensor deployment and landscape variability.
This research established an operational method for flood mapping using both free and commercial satellite imagery, which is supported by data analysis. The proposed approach determined the best sensor configurations for different land cover environments and demonstrated the significance of texture elements and classifier optimization for better flood identification. The future development should include integrating new high-resolution and real-time datasets while creating automatic hyperparameter selection systems and developing deep learning models for active flood surveillance. The framework requires additional testing across diverse geographic regions and climate zones to establish its universal application in disaster management planning and climate resilience development.