Inundated Vegetation Mapping Using SAR Data: A Comparison of Polarization Configurations of UAVSAR L‑Band and Sentinel C‑Band

: Flood events have become intense and more frequent due to heavy rainfall and hurricanes caused by global warming. Accurate floodwater extent maps are essential information sources for emergency management agencies and flood relief programs to direct their resources to the most affected areas. Synthetic Aperture Radar (SAR) data are superior to optical data for floodwater map‑ ping, especially in vegetated areas and in forests that are adjacent to urban areas and critical in‑ frastructures. Investigating floodwater mapping with various available SAR sensors and comparing their performance allows the identification of suitable SAR sensors that can be used to map inundated areas in different land covers, such as forests and vegetated areas. In this study, we investigated the performance of polarization configurations for flood boundary delineation in vegetated and open areas derived from Sentinel1b, C‑band, and Uninhabited Aerial Vehicle Synthetic Aperture Radar (UAVSAR) L‑band data collected during flood events resulting from Hurricane Florence in the east‑ ern area of North Carolina. The datasets from the sensors for the flooding event collected on the same day and same study area were processed and classified for five landcover classes using a ma‑ chine learning method—the Random Forest classification algorithm. We compared the classification results of linear, dual, and full polarizations of the SAR datasets. The L‑band fully polarized data classification achieved the highest accuracy for flood mapping as the decomposition of fully polar‑ ized SAR data allows land cover features to be identified based on their scattering mechanisms.


Introduction
Flooding is one of the most frequent natural disasters that causes damage to properties, destruction to crops, and even death to humans and animals every year worldwide.In 2016, flooding affected more than 74 million people worldwide, causing 4720 deaths and more than $57 million in economic losses [1].Global climate change causes extreme weather events that can increase flood intensity and frequency [2,3].As climate change continues to rise, many urban and rural areas will be prone to frequent floods, especially in low-elevated coastal areas [4].Rapid flood extent mapping helps flood relief programs and emergency management agencies to direct their resources to the most affected areas [5][6][7][8].
Floodwater extent maps have been generated from field surveys to hydrodynamic modeling to remotely sensed data [9][10][11][12][13][14].Delineating flooded areas using field data is usually expensive and impractical [15].Many researchers have used satellite and aerial imagery to generate floodwater extent maps.For example, [7] used optical remotely sensed data with high-resolution terrain data and social media photographs to produce rapid floodwater maps.Ban et al. analyzed optical data acquired by Terra Satellite to map floodwater along the Pampanga River in the Philippines and the Poyang and Dongting Lakes in China [16].Despite the successful utilization of optical data for mapping open floodwater extent, it is challenging to map floodwater in forests and vegetated areas since optical sensors cannot see through clouds and underneath vegetation.Hashemi-Beni and Gebrehiwot developed an integrated method using deep learning and region growing to map the flood extent in a vegetated area using optical data [17].While the method is promising, the quality of the floodwater map heavily relies on the quality of the topography data of the study area.Synthetic Aperture Radar (SAR) can monitor the earth's surface day and night, independent of sunlight for illumination and in all weather conditions [18].Due to the SAR signal capability of penetrating clouds, rain haze, and vegetation canopy, SAR provides valuable data that can be analyzed to generate near-real-time flood extent maps [19,20].An open, calm water surface appears dark in a SAR image due to the specular reflection property of the flat water surface that reflects the SAR signal away from the SAR sensor.The presence of water underneath vegetation enhances the SAR backscattered signal due to double-and multi-bounce effects between water and vegetation structures such as trunks and stems [21].Therefore, flooded vegetation appears bright in a SAR image.
Recently, many satellites carry onboard SAR sensors such as PALSAR-2, Sentinel-1, Radarsat-2, TerraSAR-X, and CosmoSkyMed.These SAR systems monitor the earth's surface using various wavelengths, including X, C, and L-bands.A longer wavelength is more suitable for detecting inundated vegetation because longer wavelengths can penetrate vegetation canopy better than shorter wavelengths [5].High temporal and spatial resolution SAR images are being used by many organizations around the world to produce nearreal-time floodwater extent maps.Utilizing SAR data from various space and airborne platforms allows the production of high-temporal-resolution floodwater extent maps, especially in challenging areas like forests and vegetated areas.Inundated vegetation mapping is critical to estimate the flood extent and avoid the unseen floods that come from these areas, protecting both human life and property [17].Before utilizing data from a SAR system, the configuration of that system, its sensor characteristics and polarization, and environmental conditions must be considered because these variables affect the reliability of floodwater extent maps [22].Analyzing the contemporaneous data of various SAR systems acquired for the same area and comparing their results to ground reference data allows the identification of the suitable SAR system, in terms of cost and accuracy, for the particular environmental condition.Airborne SAR imagery, such as UAVSAR, can provide high spatial and temporal resolution data to monitor flooded areas in detail [5].
In contrast, satellite SAR data such as Sentinel-1b (S-1b) allows the monitoring of floodwater extent over a larger area [21].In this study, we investigated and compared the performance of two SAR configurations (C-band and L-band) using different polarizations to detect and map floodwater extent in varying land cover types, including dense forests, short vegetation, and open areas over a flood-prone study area.To make the analyses comparable, which is a significant issue in the field of RS data analytics, and to identify promising strategies, the datasets from two important SAR systems (Sentinel 1b and NASA UAVSAR) were analyzed for the same flooding event in our study area due to Hurricane Florence in 2018.

Study Area and Data
The study area was located in the eastern North Carolina coastal plain area.This area is a flood-prone area covered with dense vegetation affected by many hurricanes causing heavy rain and flooding in many counties in the state.In 2018, Hurricane Florence produced high amounts of precipitation, causing devastating flooding.Fifty-three people have been confirmed dead due to Hurricane Florence, and 77% of the causalities were in N.C.'s rural flood plain [23].In addition, Florence was classified as one of the top 10 costliest hurricanes in United States history.As a result, FEMA identified 44 counties in the state of N.C. for federal disaster assistance.The study area is shown in Figure 1, where both UAVSAR data and S-1b SAR acquired on the same day are available.
have been confirmed dead due to Hurricane Florence, and 77% of the causalities were in N.C.'s rural flood plain [23].In addition, Florence was classified as one of the top 10 costliest hurricanes in United States history.As a result, FEMA identified 44 counties in the state of N.C. for federal disaster assistance.The study area is shown in Figure 1, where both UAVSAR data and S-1b SAR acquired on the same day are available.UAVSAR data and Sentinel-1b SAR data acquired on the same date were downloaded from the Alaska Satellite Facility for the study.

•
Auxiliary data: Permanent surface water data were downloaded from the U.S. Fish and Wildlife Services National Wetland Inventory.These data were used to separate permanent open water from temporarily flooded open water and to identify permanently inundated vegetation from temporarily inundated vegetation.In addition, the UAVSAR data and Sentinel-1b SAR data acquired on the same date were downloaded from the Alaska Satellite Facility for the study.

•
Auxiliary data: Permanent surface water data were downloaded from the U.S. Fish and Wildlife Services National Wetland Inventory.These data were used to separate permanent open water from temporarily flooded open water and to identify permanently inundated vegetation from temporarily inundated vegetation.In addition, the research used UAV optical data to generate training and validation samples by visually identifying land cover classes.The UAV optical datasets were obtained from the National Oceanic and Atmospheric Administration.

Data Preprocessing
SAR backscattered energy is influenced by terrain variation, SAR acquisition geometry, and noise.Therefore, some data preprocessing was conducted to improve SAR data quality by reducing noise and correcting the distortions due to SAR acquisition geometry and topography variations.In addition, the preprocessing ensures that variation in SAR backscatter values is only influenced by land cover variation.Therefore, SAR preprocessing is essential for accurate land cover classifications.

Sentinel-1b SAR Image
The preprocessing framework of Sentinel-1b data includes (a) orbit file update, (b) border noise removal, (c) radiometric calibration, (d) speckle removal, and (f) geometric and radiometric terrain correction.
(a) Orbit file update: The orbit vector information in the metadata associated with the SAR product is created using an onboard navigation solution.The precise orbit information, including the satellite velocity and position during the image acquisition, is available by Copernicus Precise Orbit Determination (POD) Service after SAR image production.Precise orbit information is needed for data geometric and radiometric calibration.
(b) Border noise removal: The GRD level 1 production from Sentinel-1 raw data involves some processing steps.These steps include sampling start time correction to compensate for the earth's curvature and azimuth and range compression.Unfortunately, these processes generate artifacts, including radiometric and no-value samples at the image borders [24], and the noise creates thin artifacts in the azimuth and range directions [25].Therefore, the border noise correction eliminates low-intensity noise and invalid data on the image edges.
(c) Radiometric calibration: This step is applied to ensure that a SAR image's pixel value directly corresponds to the radar backscatter of the sensed area.The calibration is essential for the quantitative use of level 1 SAR data.Radiometric calibration converts digital numbers in SAR image pixels to sigma naught (σ0) which is sensitive to the properties of the scattering area, incident angle, polarization, and radar wavelength [24].The scattering coefficient (σ0) is the measure of radar-returned signal strength compared to the expected signal strength reflected by a horizontal area of a one-meter square.Sigma naught is dimensionless, usually expressed in decibels (dB), and generally computed using the pixel's digital number (DN) and SAR sensor calibration factor (K) [26]: (d) Speckle removal: Speckles are noise that have a pepper appearance in SAR images.Speckles appear as the result of coherence interference between signals reflected by many elements within a resolution cell.Lee filter applies local statistics, mean, and variance of sample pixels and uses a directional window with lower mean square error [27].In addition, the Lee filter keeps the structure in the SAR image, preserving edges and characteristic patterns associated with temporally flooded areas [26].To remove speckles, we applied a Lee filter with a window dimension of 7 by 7.
(e) Geometric and radiometric terrain correction: SAR sensor tilt and the variation in terrain height can cause geometric distortion in SAR images.We applied the range doppler terrain correction to remove geometric terrain distortion.This method uses orbit vector information, the radar timing information in the annotation file, the slant-to-ground range conversion factor, and a reference digital elevation model (DEM) to geocode SAR images [28].SAR image geocoding allows the location of land cover features, combining SAR data acquired by different platforms and acquisition geometry.Georeferencing is the process of rectifying SAR images to a map projection and DEM to remove distortion in SAR images caused by topography [28].The result of the geometric correction is a geocoded SAR image with a similar geometrical representation of the sensed area.The local incident angle and topography variation influence the radar-reflected signal [26].The terrain variations affect the brightness of the radar's received energy which can reduce the accuracy of land cover classification [29].The radiometric calibration implemented here was proposed by [30]: where I is the intensity, DN is the digital number of the image, and K is the calibration constant.Then, radiometric normalization is applied using: where θ is the local incident angle, and θref is the incident angle between SAR incoming ray and normal to reference ellipsoid model of the earth's surface.

UAVSAR Data
The UAVSAR processing includes (a) extraction of the coherency matrix (T3) to identify physical properties of the scattering surface, (b) speckle noise removal, (c) polarization orientation compensation, (d) T3 matrix decomposition, and (e) radiometric terrain correction and incidence angle variation effects correction.
A backscattered polarized signal from a target contains information about the target geometry, orientation, and geophysical properties.When a scattering surface reflects a SAR wave, the physical properties of the scattering surface can be extracted from the coherence matrix of the backscattered signal.The coherency matrix T represents each pixel in the polarimetric SAR image.The coherence matrix is a 3-by-3 nonnegative definite Hermitian matrix [31]: The coherence matrix represents the contribution of three scattering mechanisms: volume, double-bounce, and surface scattering.Comparing backscatter contributions allows retrieval of relative contribution weights of each scattering component [32].T = PsT surface + PdT double bounce + PvT volume (5) where Ps, Pd, and Pv denote the power of each scattering mechanism [33].
(b) Speckle removal: Speckles are noises that appear in SAR images due to coherence interference of reflected waves from many backscatter points in the same pixel [34].Speckles can reduce SAR image segmentation and classification accuracy [35].We applied a 7-by-7 refined Lee filter.This filter preserves polarimetric properties and statistical correlation between channel features, edge sharpness, and point targets.The refined Lee filter applies local statistics using an edge-aligned window.Applying the Lee filter can significantly improve polarimetric SAR image classification performance [33].
(c) Polarization orientation compensation: The polarization orientation angle (POA) refers to the angle between the major axis of the polarization ellipse and the horizontal axis [35].The POA of the scattering surface affects polarimetric SAR signature, and without the application of orientation compensation, decomposition models may produce incorrect scattering characteristics leading to misclassification.The polarization de-orientation compensates overestimation of volume scattering (especially in forest areas) that may occur due to the effects of terrain slope on the radar-backscattered signal [5].
(d) T3 coherency matrix decomposition: Decomposition techniques extract scattering mechanisms of the earth's surface features in PolSAR data.Decomposition methods decompose total power in PolSAR image pixels into different scattering mechanisms [36].Dominant scattering mechanisms can be identified by decomposing the T3 matrix because polarimetric SAR data are influenced by target physical properties and geometry [5].We applied the Freeman and Durden polarimetry decompositions model.This model identifies and isolates three radar scattering mechanisms: volume scattering, double-bounce scattering, and single or odd scattering mechanisms.The result of this method is three image files representing portions of radar backscatter energy associated with each of the radar backscattering mechanisms.
(e) Radiometric terrain correction and incident angle variation effects correction: As mentioned above, terrain variation affects the brightness of SAR images even in areas with the same land cover class.Land cover signatures often become difficult to recover without terrain radiometric correction, especially in SAR image foreshortening, shadows, and layover pixels [37].The local incident angle varies from 22 degrees in the near range to 67 degrees in the far range.This incident angle variation from near range to far range generates illumination gradients across SAR image swaths.Radar backscattered values vary with the variation in incident angle and land cover features.The UAVSAR data were radiometrically corrected for terrain and incident angle variation effects.Backscatter signatures can be misjudged without these corrections, leading to less accurate land cover classification.

Data Classification
Image classification techniques are the process of extracting information from an image to identify land cover features based on pixel values [38].Based on the level at which image classification is performed, image classification techniques are divided into pixellevel methods and object-level methods.Pixel-level methods consider each pixel belonging to a single land cover class based on either unsupervised classification algorithms or supervised classification algorithms.The unsupervised algorithms do not require training data or prior knowledge of the study area and can identify land cover classes by clustering image values [39].In contrast, supervised algorithms learn from training samples and label a pixel's class by comparing its spectral information to training samples [40].Object-level classification methods apply image segmentation techniques to divide remote sensing images into objects or geographic features and label them based on the spectral properties and other relevant measures [40].
In this study, we applied the Random Forest (RF) supervised classification algorithm.The RF classification method is fast and robust to outliers and noise [41].RF classifiers can handle large datasets with many variables and identify important variables in the classification process [42].RF is an ensemble classifier that generates many decision trees, and its final prediction is based on the majority of votes obtained from all trees [43].The RF classifier builds many classification and regression trees from training data using random sampling with replacement [44].This indicates that the same training sample can be drawn more than once, while some samples may not be drawn at all [45].Using a random feature selection, trees are grown on a new training set without pruning [41].Each decision tree is generated independently, and it is up to the analyst to define the number of trees and features to split each node.The RF algorithm generates decision trees with low bias and high variances [45].One-third of the bootstrap training set is used for internal validation to evaluate the performance of the RF model [41].It is characterized by high prediction accuracy and fast computation process [40] Generally, decision-tree-based algorithms apply many approaches, including gain ratio, Gini index, and chi-square, to identify a suitable attribute that results in maximum dissimilarity between classes [42].RF classifiers usually compute the Gini index to identify the characteristic that yields the best split option by measuring the impurity of the characteristics with respect to other classes [42].Gini index is computed using the following: where T is the training sample and f (Ci, T)/|T| represents the probability of selecting pixels that belong to class Ci.In RF classification algorithms, decision trees are extended to their full depth without pruning using a combination of variables or features.As a result, the RF classifier is more advanced than other decision-based methods because the performance of tree-based classification methods is affected by the choice of pruning approaches and not attribute selection measures [43].The computation time needed to build the RF classification model can be computed using the following [41]: where T represents the number of trees, M is the number of features or variables used in node split, and N is the size of training samples.The classification process includes (a) land cover sample labeling and (b) training, prediction, and validation.
(a) Land cover sample labeling The UAV high-resolution optical data were overlaid with SAR data.We visually identified five different land cover types: open water, bare ground, grassland, inundated vegetation, and non-inundated vegetation.Over 120 training samples for each of the five land covers were manually annotated.
(b) Training, prediction, and validation Supervised classification methods require training samples selected from sites with known land cover types to represent land cover classes in the area of interest.Supervised classifiers compare the spectral information of image pixels with the spectral pattern in the training samples.Then, some statistical and classification decision rules are used to assign each image pixel to a class with the highest similarity to the pixel [40].Ground truth data are generally used to assess the accuracy of the classifier.In this study, we randomly selected 75% of the training sample pixels for training, and 25% of the samples were withheld for validation.This process generated 214,572 point samples for training and 7143 points for validation for the S-1b data.For The UAVSAR data, 51,980 and 17,182 point samples were randomly selected for training and validation of the predictions.The RF classifier was trained using the selected training sample pixels for each land cover type.Various numbers of decision trees and variables per split were tested.The model with 300 trees and 3 variables per split has achieved high training accuracy, so we classified the data using this model.

Results
The classification accuracy was estimated using validation samples, i.e., the annotated data that had not been seen by the RF classifier.

UAVSAR L-Band Classification Results
The UAVSAR data was classified using three models: linear polarization (VV/VH), dual polarization, and full polarization.
(a) L-band linear polarization: The linear polarizations VV and VH were analyzed and segmented using the RF method to create a flood map of the study areas.The performance of the classification results was evaluated using a confusion matrix representing counts from predicted and actual values in the classifications.The confusion matrix of this VV model is shown in Table 1, which was generated by comparing the land cover classification results obtained from this model to the training samples.After generating the confusion matrix, we computed the training accuracy using: Overall training Accuracy (OA) = total number o f correctly classi f ied pixels total number o f pixels in the traing sample OA f or the VV Model = 22,124 22,346 = 99.01% The producer and consumer accuracy for the UAVSAR VV model is shown in Figure 2. The classification accuracy was also evaluated using kappa statistics.Kappa statistics provide additional accuracy assessment measuring test interrater reliability.Kappa statistic (k) is calculated by subtracting the summation of chance agreement from the summation of the agreement of the results [44].
where P o is observed agreement, and Pch is the chance agreement.K for UAV training: The producer and consumer accuracy for the UAVSAR VV model is shown in Figure 2. The classification accuracy was also evaluated using kappa statistics.Kappa statistics provide additional accuracy assessment measuring test interrater reliability.Kappa statistic (k) is calculated by subtracting the summation of chance agreement from the summation of the agreement of the results [44].
where Po is observed agreement, and Pch is the chance agreement.K for UAV training:   The validation error matrix for the classification of the UAVSAR VV polarization data was generated by comparing the classification results with the reference data samples that the classifier has not seen, hence providing a reliable accuracy assessment.The overall validation accuracy for this model is 92.35%, with kappa statistics of 83.65%.The validation error matrix for this model is shown in Table 2.
The linear polarization of VH was also analyzed and segmented using the RF method to create a flood map of the study areas.The results of the classification are described in Figures 3 and 4. The validation error matrix for the classification of the UAVSAR VV polarization data was generated by comparing the classification results with the reference data samples that the classifier has not seen, hence providing a reliable accuracy assessment.The overall validation accuracy for this model is 92.35%, with kappa statistics of 83.65%.The validation error matrix for this model is shown in Table 2.The linear polarization of VH was also analyzed and segmented using the RF method to create a flood map of the study areas.The results of the classification are described in Figures 3 and 4

(b) L-band dual polarization:
To study the performance of the UAVSAR dual polarization for flood mapping, a composite image was created using the VV, VH, and the ratio |VV|/|VH|.This image was classified using the RF classification algorithm.Figure 5      (c) L-band full polarization: This study also utilized the UAVSAR full polarization for flood mapping.The performance of the UAVSAR data classification is described in Figure 7 and Figure 8. (c) L-band full polarization: This study also utilized the UAVSAR full polarization for flood mapping.The performance of the UAVSAR data classification is described in Figures 7 and 8.
The overall training accuracy and validation accuracy for the UAVSAR decomposition model are 99.45% and 98.68%, respectively.The kappa statistic for the validation reference samples is 97.47%.The overall training accuracy and validation accuracy for the UAVSAR decomposition model are 99.45% and 98.68%, respectively.The kappa statistic for the validation reference samples is 97.47%.

Sentinel C-band Classification Results
We also generated the confusion matrices for the classification results obtained from S-1b linear polarization (i.e., VV and VH) and dual polarization by comparing the classifier outputs to training samples.
(a) C-band linear polarization: The linear polarizations from sentinel C-band VV-and VH-polarized data were analyzed and segmented using the RF method to create a flood map of the study areas.The performance of the classification results was evaluated using a confusion matrix representing counts from predicted and actual values in the

Sentinel C-Band Classification Results
We also generated the confusion matrices for the classification results obtained from S-1b linear polarization (i.e., VV and VH) and dual polarization by comparing the classifier outputs to training samples.C-band dual polarization: Similar to the UAVSAR dual polarization analysis for flood mapping, a composite image was created using the Sentinel C-band VV, VH, and the ratio |VV|/|VH|.This image was classified using the RF classification algorithm.Figure 12 shows the classification performance for the training and validation of the C-band dualpolarized data for flood mapping.This model has an overall training accuracy of 97.53% with a kappa coefficient of 94.61%.The validation error samples for this model are shown in Figure 12c.This model has an overall validation accuracy of 91.96%, with kappa statistics of 82.01%(Figure 13).C-band dual polarization: Similar to the UAVSAR dual polarization analysis for flood mapping, a composite image was created using the Sentinel C-band VV, VH, and the ratio |VV|/|VH|.This image was classified using the RF classification algorithm.Figure 12 shows the classification performance for the training and validation of the C-band dualpolarized data for flood mapping.This model has an overall training accuracy of 97.53% with a kappa coefficient of 94.61%.The validation error samples for this model are shown in Figure 12c.This model has an overall validation accuracy of 91.96%, with kappa statistics of 82.01%(Figure 13).We noticed that both Sentinel and UAVSAR datasets accurately predicted open water areas as expected, because of the high contrast between an open water surface and other land cover features in the study area.This high contrast results from the specular reflection property of calm open water surfaces, causing radar signals to be reflected away from the SAR antenna and making pixels of open water appear dark in the SAR datasets.In addition, some inundated vegetated areas that appeared bright in the SAR images were also accurately detected.The presence of water underneath vegetation enhanced the SAR backscattered signal due to double-bounce and multi-bounce effects between flat water surfaces and vertical vegetation structures, such as trunks and stems, leading to the higher prediction accuracy of inundated vegetation pixels.
The results show that co-polarization VV performed slightly better than the crosspolarization VH for floodwater mapping using both L-band and C-band.As expected, the combination of L-band SAR polarizations (dual and full polarization) performed better than retrieving the floodwater class from a single polarization.However, the classification of the S-1b C-band dual polarization did not increase the flood map accuracy more than the results obtained from the co-polarization of VV.Because the L-band penetrates the vegetation canopy better than the C-band, the overall accuracy of the L-band dual-polarized data was higher than that of the C-band dual-polarized data.The L-band fully polarized data classification achieved the highest accuracy for flood mapping because the decomposition of fully polarized SAR data allows land cover features to be identified based on their scattering mechanisms.    .

Conclusions
Inundated vegetation mapping is critical to estimate the flood extent and avoid the unseen floods that come from these areas, thus protecting both human life and property.This The backscattered signal enhancement due to double-and multi-bounce effects between the flat water surface and vertical vegetation structures helped identify inundated vegetation with high accuracy.The S-1b VV model had an overall accuracy of 92.57%, while the S-1bVH model achieved 86.97% overall classification accuracy.Therefore, sentinel data showed promising performance in predicting inundated vegetation in the area.Our results showed that the S-1b co-polarization performed slightly better than the S-1b crosspolarization, which agrees with previous studies [46][47][48].However, our results showed that the S-1b C-band dual polarization did not increase the flood map accuracy more than the results obtained from the co-polarization VV.Therefore, the VV-polarized data from Sentinel 1b are suitable for mapping inundated vegetation in large study areas with reasonable accuracy.In addition, satellite data allow the monitoring of regional floodwater mapping with less cost than airborne SAR datasets.

Figure 1 .
Figure 1.The study area: SAR images of two subareas during Hurricane Florence.• UAVSAR data: The dataset includes high-resolution SAR data captured over several flight lines with repeated path observations from September 18 to 23 September 2018.UAVSAR is a fully polarized SAR system that collects data using L-band.In this study, we analyzed the ground-range-projected (GRD) format acquired over flight line 31509.These data have a pixel spacing of 6 m by 5 m.• Sentinel-1b: The Sentinel-1 mission contains two identical satellites: Sentinel-1a and Sentinel-1b.Each satellite provides global coverage with a revisit time of 12 days.By utilizing data from the two satellites, the global coverage temporal resolution can increase to 6 days.Sentinel-1 acquires data over the land surface using the Interferometric Wide-Swath mode with 250-km swath width and 5 × 20-m spatial resolution.Sentinel-1b collects data using dual polarization C-band (5.405 GHz, 5.6 cm wavelength), i.e., VV (vertical sent-vertical received) and VH (vertical sent-horizontal received).This research used Sentinel-1b SAR level 1 acquired on 19 September 2018 from path 77 using VV and VH polarizations.

Figure 1 .
Figure 1.The study area: SAR images of two subareas during Hurricane Florence.• UAVSAR data: The dataset includes high-resolution SAR data captured over several flight lines with repeated path observations from September 18 to 23 September 2018.UAVSAR is a fully polarized SAR system that collects data using L-band.In this study, we analyzed the ground-range-projected (GRD) format acquired over flight line 31509.These data have a pixel spacing of 6 m by 5 m.• Sentinel-1b: The Sentinel-1 mission contains two identical satellites: Sentinel-1a and Sentinel-1b.Each satellite provides global coverage with a revisit time of 12 days.By utilizing data from the two satellites, the global coverage temporal resolution can increase to 6 days.Sentinel-1 acquires data over the land surface using the Interferometric Wide-Swath mode with 250-km swath width and 5 × 20-m spatial resolution.Sentinel-1b collects data using dual polarization C-band (5.405 GHz, 5.6 cm wavelength), i.e., VV (vertical sent-vertical received) and VH (vertical sent-horizontal received).This research used Sentinel-1b SAR level 1 acquired on 19 September 2018 from path 77 using VV and VH polarizations.

Figure 2 .
Figure 2. Producer and consumer accuracy for L-band VV polarization.Figure 2. Producer and consumer accuracy for L-band VV polarization.

Figure 2 .
Figure 2. Producer and consumer accuracy for L-band VV polarization.Figure 2. Producer and consumer accuracy for L-band VV polarization. .

20 Figure 3 .
Figure 3.The accuracy assessment of the UAVSAR L-band VH-polarized data for flood mapping: (a) confusion matrix for training the RF model, (b) producer and consumer accuracy for training, (c) validation error matrix for the classification, and (d) producer and consumer accuracy for validation.

Figure 3 .
Figure 3.The accuracy assessment of the UAVSAR L-band VH-polarized data for flood mapping: (a) confusion matrix for training the RF model, (b) producer and consumer accuracy for training, (c) validation error matrix for the classification, and (d) producer and consumer accuracy for validation.

Figure 3 .
Figure 3.The accuracy assessment of the UAVSAR L-band VH-polarized data for flood mapping: (a) confusion matrix for training the RF model, (b) producer and consumer accuracy for training, (c) validation error matrix for the classification, and (d) producer and consumer accuracy for validation.

Figure 4 .
Figure 4. Classification results obtained from the UAVSAR L-band VV (a) and VH (b) polarized data.
describes the classification performance for the training and validation of the L-band dual polarization for flood mapping.This model has an overall training accuracy of 99.61% with a kappa coefficient of 99.18.The validation error samples for this model are shown in Figure 5 and Figure 6.This model shows overall validation accuracy of 94.95%, with kappa statistics of 89.13%.

Figure 4 . 20 Figure 5 .
Figure 4. Classification results obtained from the UAVSAR L-band VV (a) and VH (b) polarized data.(b) L-band dual polarization: To study the performance of the UAVSAR dual polarization for flood mapping, a composite image was created using the VV, VH, and the ratio |VV|/|VH|.This image was classified using the RF classification algorithm.Figure 5 describes the classification performance for the training and validation of the L-band dual polarization for flood mapping.This model has an overall training accuracy of 99.61% with a kappa coefficient of 99.18.The validation error samples for this model are shown in Figures 5 and 6.This model shows overall validation accuracy of 94.95%, with kappa statistics of 89.13%.Remote Sens. 2022, 14, x FOR PEER REVIEW 11 of 20

Figure 5 .
Figure 5.The accuracy assessment of the UAVSAR L-band dual-polarized data for flood mapping: (a) confusion matrix for training the RF model, (b) producer and consumer accuracy, and (c) validation error matrix for the classification.

Figure 5 .
Figure 5.The accuracy assessment of the UAVSAR L-band dual-polarized data for flood mapping: (a) confusion matrix for training the RF model, (b) producer and consumer accuracy, and (c) validation error matrix for the classification.

Figure 6 .
Figure 6.Classification results obtained from the UAVSAR L-band dual-polarized data.

Figure 6 .
Figure 6.Classification results obtained from the UAVSAR L-band dual-polarized data.

Figure 7 .
Figure 7.The accuracy assessment of the UAVSAR L-band full-polarized data for flood mapping: (a) confusion matrix for training the RF model, (b) producer and consumer accuracy, and (c) validation error matrix for the classification.

Figure 7 .
Figure 7.The accuracy assessment of the UAVSAR L-band full-polarized data for flood mapping: (a) confusion matrix for training the RF model, (b) producer and consumer accuracy, and (c) validation error matrix for the classification.

Figure 7 .
Figure 7.The accuracy assessment of the UAVSAR L-band full-polarized data for flood mapping: (a) confusion matrix for training the RF model, (b) producer and consumer accuracy, and (c) validation error matrix for the classification.The overall training accuracy and validation accuracy for the UAVSAR decomposition model are 99.45% and 98.68%, respectively.The kappa statistic for the validation reference samples is 97.47%.

Figure 8 .
Figure 8. Classification results obtained from the UAVSAR L-band full-polarized data.

Figure 8 .
Figure 8. Classification results obtained from the UAVSAR L-band full-polarized data.
(a) C-band linear polarization: The linear polarizations from sentinel C-band VV-and VH-polarized data were analyzed and segmented using the RF method to create a flood map of the study areas.The performance of the classification results was evaluated using a confusion matrix representing counts from predicted and actual values in the classifications.The performance of the S-1b VV and VH data classification is described in Figures 9 and 10.Remote Sens. 2022, 14, x FOR PEER REVIEW 13 of 20 classifications.The performance of the S-1b VV and VH data classification is described in Figures 9 and 10 .

Figure 9 .
Figure 9.The accuracy assessment of the Sentinel C-band VV-polarized data for flood mapping: (a) confusion matrix for training the RF model, (b) producer and consumer accuracy, and (c) validation error matrix for the classification.

Figure 9 .
Figure 9.The accuracy assessment of the Sentinel C-band VV-polarized data for flood mapping: (a) confusion matrix for training the RF model, (b) producer and consumer accuracy, and (c) validation error matrix for the classification.

Figure 9 .
Figure 9.The accuracy assessment of the Sentinel C-band VV-polarized data for flood mapping: (a) confusion matrix for training the RF model, (b) producer and consumer accuracy, and (c) validation error matrix for the classification.

Figure 10 .
Figure 10.The accuracy assessment of the Sentinel C-band VH-polarized data for flood mapping: (a) confusion matrix for training the RF model, b) producer and consumer accuracy, and (c) validation error matrix for the classification.The overall training accuracy for the S-1bVVSAR model is 93.52% (17,238/18,432).The kappa statistic for the overall training accuracy is 87.81%.The VV model producer and consumer accuracy are shown in Figure 9b.We noticed that the model predicted open water and inundated vegetation very well.However, for non-inundated vegetation, the model showed low producer accuracy.The validation error matrix for the S-1b (VV) model is shown in Figure 10 (c) .The overall validation accuracy for the S-1b SAR VV model is 92.57% (7437/8034), and the kappa statistic for the validation accuracy is 85.81%.

Figure 10 .Figure 11 .
Figure 10.The accuracy assessment of the Sentinel C-band VH-polarized data for flood mapping: (a) confusion matrix for training the RF model, (b) producer and consumer accuracy, and (c) validation error matrix for the classification.The overall training accuracy for the S-1bVVSAR model is 93.52% (17,238/18,432).The kappa statistic for the overall training accuracy is 87.81%.The VV model producer and consumer accuracy are shown in Figure 9b.We noticed that the model predicted open water and inundated vegetation very well.However, for non-inundated vegetation, the model showed low producer accuracy.The validation error matrix for the S-1b (VV) model is shown in Figure 10c.The overall validation accuracy for the S-1b SAR VV model is 92.57% (7437/8034), and the kappa statistic for the validation accuracy is 85.81%.The overall accuracy for the S-1b SAR data (VH polarization) is 86.97% (6827/7850).This model also predicted open water and inundated vegetation with high accuracy (Figure 11).The S-1b VH model overall validation accuracy had a kappa statistic of 75.2%.Remote Sens. 2022, 14, x FOR PEER REVIEW 14 of 20

Figure 11 .
Figure 11.Classification results obtained from the Sentinel C-band VV-(a) and VH-polarized (b) data.

Figure 11 .
Figure 11.Classification results obtained from the Sentinel C-band VV-(a) and VH-polarized (b) data.

Figure 12 .
Figure 12.The accuracy assessment of the Sentinel C-band dual-polarized data for flood mapping: (a) confusion matrix for training the RF model, (b) producer and consumer accuracy, and (c) validation error matrix for the classification.

Figure 12 . 20 Figure 13 .
Figure 12.The accuracy assessment of the Sentinel C-band dual-polarized data for flood mapping: (a) confusion matrix for training the RF model, (b) producer and consumer accuracy, and (c) validation error matrix for the classification.Remote Sens. 2022, 14, x FOR PEER REVIEW 15 of 20

Figure 13 .
Figure 13.Classification results obtained from the Sentinel C-band dual-polarized data.
Finally, the flooded vegetation and open floodwater areas were identified by masking permanent water areas.The permanent water layers over our study area, including water under vegetation and open water, were obtained from US fish and wildlife services.The floodwater, including both inundated vegetation and open floodwater detected by UAVSAR and S-1b SAR dual-polarized datasets, are shown in Figures 14 and 15, respectively.
study investigated the performance of different polarization configurations of UAVSAR Lband and Sentinel C-band datasets for inundated vegetation mapping over a flood-prone study area in North Carolina (USA) during Hurricane Florence.Both datasets were suitable for predicting open water with high accuracy.The forward reflection characteristics associated with calm, open water bodies result in low radar backscatter values causing high contrast between open water and other land cover features, leading to the accurate identification of pixels of open water areas.However, inundated vegetation pixels exhibit the double-bounce scattering mechanism identified when decomposing the coherence matrix of fully polarized L-band SAR data.This helps predict flooded vegetation with high accuracy since L-band signals penetrate vegetation canopies more than C-band.The overall floodwater classification accuracy obtained from the Freeman-Durden decomposition of fully polarized L-band reached 98.68%.In addition, the Freeman-Durden decomposition model of the fully polarized SAR data identifies different scattering mechanisms corresponding to different land cover classes, improving the accuracy of the land cover classification.

Table 1 .
Confusion matrix for VV polarization of UAVSAR L-band.

Table 2 .
Validation error matrix, and producer and consumer accuracy for VV polarization of UAVSAR L-band.

Table 2 .
Validation error matrix, and producer and consumer accuracy for VV polarization of UAVSAR L-band.

Table 3
summarizes the training and validation accuracy of the RF classification of the UAVSAR L-band and Sentinel 1b C-band for flood mapping using different polarization configurations.

Table 3 .
The overal accuracy and kappa statistics for the classification results.