Next Article in Journal
Recent Advances in Aptamer-Based Point-of-Care Testing
Previous Article in Journal
A Robust Rotation-Equivariant Feature Extraction Framework for Ground Texture-Based Visual Localization
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Closing the Domain Gap: Can Pseudo-Labels from Synthetic UAV Data Enable Real-World Flood Segmentation?

by
Georgios Simantiris
1,*,
Konstantinos Bacharidis
1,2 and
Costas Panagiotakis
1,2,*
1
Department of Management Science and Technology, Hellenic Mediterranean University, 72100 Agios Nikolaos, Greece
2
Institute of Computer Science, FORTH (Foundation for Research & Technology–Hellas), 70013 Heraklion, Greece
*
Authors to whom correspondence should be addressed.
Sensors 2025, 25(12), 3586; https://doi.org/10.3390/s25123586
Submission received: 17 April 2025 / Revised: 31 May 2025 / Accepted: 4 June 2025 / Published: 6 June 2025

Abstract

:
We present a novel methodology for generating and filtering synthetic Unmanned Aerial Vehicle (UAV) flood imagery to enhance the generalization capabilities of segmentation models. Our framework combines text-to-image synthesis and image inpainting, using curated prompts and real-world segmentation masks to produce diverse and realistic flood scenes. To overcome the lack of human annotations, we employ an unsupervised pseudo-labeling method that generates segmentation masks based on floodwater appearance characteristics. We further introduce a filtering stage based on outlier detection in feature space to improve the realism of the synthetic dataset. Experimental results on five state-of-the-art flood segmentation models show that synthetic data can closely match real data in training performance, and combining both sources improves model robustness by 1–7%. Finally, we investigate the impact of prompt design on the visual fidelity of generated images and provide qualitative and quantitative evidence of distributional similarity between real and synthetic data.

Graphical Abstract

1. Introduction

Natural disasters have long affected humanity, with climate change intensifying their frequency and severity. These events cause significant loss of life, property damage, and service disruptions, such as electricity and transportation, while posing serious health risks. The economic and psychological toll is profound [1].
Technological advances such as Pattern Recognition (PR), Deep Learning (DL), Machine Learning (ML), and Artificial Intelligence (AI) provide powerful tools for disaster detection, risk reduction, and response management. As reviewed in [2,3,4], these technologies hold great promise for future disaster response, especially AI and ML in the domain of computer vision (CV), through the use of predictive models that analyze large datasets, identify patterns, forecast potential disasters, and provide early warnings of hazardous events [4,5]. Deep learning (DL) is increasingly used for flood detection and segmentation, overcoming the limitations of traditional mapping [6]. However, DL models require extensive labeled data, which is difficult to obtain in disaster scenarios, requires multiple experts, is time-consuming, and is prone to human annotation.
To provide a comprehensive overview of the available resources in flood-related visual understanding from aerial imagery, we compiled a selection of ten publicly available datasets, namely AIDER [7], ISBDA [8,9], FloodNet [10], FAD [11], Spacenet-8 [12], FSSD [13], WaterBodies [14], Incidents1M [15], RescueNet [16], and BlessemFlood21 [17] released between 2019 and 2024. These are illustrated in Figure 1, which depicts a bubble chart where each bubble represents a dataset. The horizontal axis indicates the year of publication, while the vertical axis corresponds to the dataset image size. The area of each bubble is scaled proportionally to the dataset size to emphasize the disparity in scale across datasets. Each bubble is also annotated with the dataset name, and a distinct color is used for visual separation. Gray bubbles indicate datasets without flood annotations (AIDER [7], ISBDA [8,9], Incidents1M [15]), and for AIDER [7] and Incidents1M [15] only the subset of flood-related aerial images is considered in the reported sizes.
The increasing trend in dataset size over recent years highlights the growing effort to support deep learning-based flood analysis and emergency response using aerial visual data. However, the creation of large-scale annotated datasets remains a significant challenge. Manual annotation is labor-intensive, often requiring the expert interpretation of complex visual patterns. Furthermore, achieving scene diversity is difficult, as the aerial imagery of floods is typically limited to specific geographical regions and events, which may reduce generalizability across diverse flood scenarios.
To overcome these challenges, one viable solution is the use of synthetic data generated through advanced generative methods, such as diffusion models or generative adversarial networks (GANs). Synthetic imagery enables the creation of diverse flooding scenarios, including rare or extreme cases, thereby improving the model’s generalizability and robustness. Moreover, in the absence of ground truth labels, pseudo-labeling through automated segmentation algorithms offers a promising alternative. These methods can generate approximate annotations at scale, significantly reducing the dependence on expert input and enabling the effective training of deep learning models in a weak or unsupervised manner.
In this paper, we solve the problem of flood segmentation in UAV imagery using synthetic generated images of flooding events instead of real ones to answer the following open question: Can pseudo-labels from synthetic UAV data enable real-world flood segmentation? In our work, we propose a framework to create synthetically generated images of flooding events using two algorithms: (i) text-to-image synthesis and (ii) image inpainting based on real segmentation masks. The generated images undergo unsupervised pseudo-labeling for flood segmentation masks, and filtering using feature embeddings and clustering to refine the synthetic dataset. We train different popular Convolutional Neural Network (CNN) and transformer architectures for flood segmentation. Our results show that these models when trained on both synthetic and real images can achieve higher segmentation accuracy. We evaluate their performance on an unseen test set and compare the results to models trained with human-annotated labels and the unsupervised pseudo-labeling approach. A schematic overview of the proposed framework is shown in Figure 2.
The contributions of this work can be summarized as listed below.
  • We introduce, to the best of our knowledge, the first scalable pipeline for the unsupervised generation of synthetic aerial flood imagery, utilizing text-to-image diffusion models guided by semantically enriched prompts. To enable segmentation training without the need for manual annotation, we integrate an unsupervised pseudo-labeling approach [18], which automatically produces segmentation masks by exploiting the distinct color characteristics of floodwater and surrounding background elements.
  • We demonstrate through extensive experiments with state-of-the-art flood segmentation models that models trained solely on filtered synthetic data achieve a performance close to real-data-trained models, with minor performance drops, and introduce an approach to combine real and synthetic data in order to boost performance.
  • We systematically examine how the structure and semantics of text prompts affect the quality and realism of the generated flood imagery, identifying factors that influence scene consistency and visual fidelity.
The remainder of this paper is structured as follows: Section 2 provides a summary of related research. Our proposed unsupervised framework is described in Section 3. Section 4 outlines the experimental framework of this study. Section 5 presents the experimental results along with an in-depth discussion. Finally, Section 6 concludes the paper and outlines directions for future research.

2. Related Work

Flood segmentation in UAV and satellite imagery: Deep learning techniques, particularly CNNs, are increasingly used for flood segmentation in remote sensing imagery, surpassing traditional methods by enabling the more accurate and efficient delineation of flooded areas and enhancing decision-making processes [6]. CNNs have shown strong capabilities in flood detection from satellite imagery by leveraging temporal variations in synthetic aperture radar (SAR) and multispectral data to differentiate between permanent water bodies and flood-affected areas [19,20]. However, their effectiveness is often constrained by the reliance on pre-disaster imagery for accurate change detection. To address uncertainty in SAR-based water segmentation, Bayesian CNNs have been proposed due to their ability to estimate both the mean and variance of model parameters, providing a probabilistic understanding of predictions [21].
U-Net variants have been widely adopted for water body segmentation and flood extent mapping tasks. For instance, in [22], a modified U-Net architecture was proposed that incorporated geomorphic features and utilized pre-processed Sentinel-1 radar imagery to achieve three-class classification. This model successfully differentiated flood water from permanent water and background. Similarly, in [12,23], it was demonstrated that lightweight U-Net configurations can offer an effective balance between accuracy, computational efficiency, and robustness. The use of transfer learning and targeted data augmentation proved essential in enabling the detection of flooded infrastructure, including roads and buildings. Furthermore, in [14], the performance of various CNN architectures for water body semantic segmentation was evaluated using high-resolution satellite and aerial imagery. The U-Net model with a MobileNet-V3 backbone, along with auxiliary features and data augmentation, achieved superior segmentation accuracy.
Benchmark experiments involving semantic segmentation have validated state-of-the-art deep learning models including XceptionNet and ENet for distinguishing floodwaters from natural water bodies, and detecting inundated roads and buildings in UAV-acquired high-resolution post-disaster imagery [10]. Also, a CNN integrated into the Deep Earth Learning, Tools, and Analysis (DELTA) framework achieved high precision and recall for water segmentation across diverse datasets [24]. In [25], a multiscale attentive decoder network (ADNet) was proposed for automatic flood identification using Sentinel-1 images. When evaluated on the Sen1Floods11 benchmark dataset, ADNet outperformed recent deep learning and threshold-based approaches.
In [26], an enhanced version of the efficient neural network (ENet) architecture was adopted for the semantic segmentation of UAV footage captured during flood events. The approach integrates atrous (dilated) separable convolutions in the encoder, enlarging the receptive field without increasing computational complexity [27], and depth-wise separable convolutions in the decoder, enabling efficient feature extraction with a reduced number of parameters. Atrous convolutions have been further utilized in disaster response scenarios to enhance the efficiency of search and rescue operations during events such as floods, high tides, and tsunamis. A notable example is FASegNet, a recently proposed CNN architecture designed for the semantic segmentation of flood- and tsunami-affected areas [28].
Transformer-based architectures have also demonstrated strong performance in the semantic segmentation of remote sensing imagery. In [29], a novel approach was introduced, employing the Swin Transformer as the backbone to enhance contextual feature representation, coupled with a densely connected feature aggregation module as the decoder. Additionally, the Bitemporal Image Transformer (BiT) model proposed in [19] showed superior performance in change detection tasks by effectively identifying and localizing regions of change between image pairs.
In [30], an interactive semantic segmentation model for multi-source UAV flood images using four prompt types was presented. A prompt encoder maps prompts into a three-channel space to lower labeling costs, while an image encoder, combining Mamba and convolution operations, extracts global features. The model further improves prompt utilization by incorporating a spatial and channel attention module with residual connections. This enables a multiscale fusion and filtering of prompt information and image features across both spatial and channel dimensions.
Several unsupervised flood segmentation methods have been developed using clustering and region-growing techniques. These include object-based K-means with region growing on SAR data [31] and UAV imagery [32], datacube-based flood mapping with probabilistic thresholds [33], tile-based histogram thresholding with contextual filters [34], and graph-based segmentation using Bayesian Markov random fields [35], all demonstrating effectiveness across various imaging sources and scenarios.
In [18], a fully unsupervised segmentation method, UFS-HT-REM, was introduced for fast and accurate flood area detection using UAV-acquired color imagery without requiring pre-disaster reference images. This framework addresses flood segmentation through a parameter-free, unsupervised image analysis pipeline that progressively eliminates non-flood regions using binary masks derived from color and edge information. Specifically, non-flood areas are excluded through mask calculations applied to each channel of the LAB color space, an RGB-based vegetation index, and edge maps from the original image. A probability map of flood presence is then generated using a weighted fusion strategy, followed by a modified hysteresis thresholding process to produce the final segmentation. The method demonstrates both high accuracy and computational efficiency, making it well suited for real-time, on-board processing during UAV operations. This methodology has been used as pseudo-label generator for an unsupervised DL approach [36], and also serves the same purpose as a module in this work.
Pseudo-Label-Based Methods for Semantic Segmentation: Pseudo-labels in image segmentation are a common strategy in semi-supervised and unsupervised learning. The idea is to use automatically generated labels, often produced by an initial model, unsupervised method, or a rule-based system to train deep learning models without requiring large amounts of manually annotated data. This approach has been widely explored in general semantic segmentation tasks, such as urban scene understanding, medical imaging, and object detection.
A recent review and analysis of various PL methods and their applications in semi-supervised semantic segmentation (SSSS) underlined that training with limited labeled data by leveraging automatically generated labels can be effective [37]. Limitations of existing pseudo-label generation by leveraging enhanced class activation maps and dual attention mechanisms to produce semantically rich labels, achieving competitive or superior performance, has been addressed in [38]. Furthermore, self-supervised learning and pseudo-label refinement were integrated in a novel weakly supervised semantic segmentation (WSSS) approach, achieving near fully supervised performance by enhancing feature representation and mitigating label noise [39]. In [40], PseudoSeg was introduced, a method that generates structured pseudo-labels for training with unlabeled or weakly labeled data, demonstrating effectiveness in both low-data and high-data regimes.
In the context of flood segmentation, pseudo-labeling is less established. A few recent studies have started exploring automatic label generation for SAR and optical imagery, particularly when annotated flood datasets are limited or unavailable. A semi-supervised learning method for flood segmentation using Sentinel-1 SAR imagery, employing a cyclical training process with an ensemble of U-Net models trained on both high-confidence hand-labeled data and generated pseudo-labels was introduced in [41]. Moreover, an unsupervised deep learning framework for water extraction from multispectral imagery, combining NDWI with a binarization algorithm to generate pseudo-labels for training, was proposed [42]. A novel WSSS framework, TFCSD, has been introduced for efficient urban flood mapping, significantly reducing manual annotation by decoupling the generation of positive and negative samples [43]. The method enhances edge delineation and stability and maintains high performance even without pre-disaster data by incorporating SAM-assisted interactive labeling [44].
In our recent work [36], we proposed a novel unsupervised deep learning approach for flood segmentation in Unmanned Aerial Vehicle (UAV) imagery, which leverages automatically generated pseudo-labels as training and validation masks, thereby eliminating the need for manually annotated ground truth data. Two widely used Convolutional Neural Network (CNN) architectures for semantic segmentation were trained under this framework. The results demonstrated that training with pseudo-labels alone can achieve performance levels comparable to those obtained using conventional ground truth annotations. Finally, in [45] a semi-supervised semantic segmentation algorithm for accurate flood delineation in SAR data was proposed. The method exhibited promising results utilizing a pseudo-label generation strategy and self-supervised teacher–student models.
Deep generative models and diffusion-based image synthesis: The performance of modern DL models is fundamentally tied to the availability of large-scale datasets. In many real-world domains, such as disaster monitoring or remote sensing, data acquisition is costly. To address this bottleneck, a promising direction is the generation of synthetic data, which can serve as a scalable and controllable alternative to real-world data collection. Synthetic data can be created through simulation, procedural generation, or by leveraging deep generative models. Among generative approaches, models are typically grouped into four main categories: variational autoencoders (VAEs), generative adversarial networks (GANs), auto-regressive models, and, more recently, diffusion models [46,47].
Diffusion models have emerged as the most widely adopted family of models, in a variety of domains such as image and text-to-image synthesis [48,49,50], image inpainting [47,51], and image-to-image translation [52,53]. These models operate by learning to denoise a sample starting from pure Gaussian noise, gradually transforming it into a realistic image. They exhibit greater training stability compared to GANs and offer fine-grained control over the generation process via conditioning mechanisms, such as textual prompts or image masks, and have proven to generate sharper images compared to VAEs, which often produce blurry outputs due to their reliance on approximate posterior inference [46]. The ability of diffusion-family models to learn strong latent representations that are aligned with textual prompts, and are able to generalize to unseen scenarios (textual prompts) generating high-fidelity images is asserted by the emergence of widely exploited models such as the DALL-E series [54,55], Stable Diffusion [47], and MidJourney, which generate high-quality, customizable images from textual descriptions. Their capacity to synthesize diverse, domain-specific imagery makes diffusion models particularly suitable for augmenting datasets in applications where labeled data is scarce or costly to produce.

3. Synthetic Dataset Generation Methodology

Our proposed approach is a framework that enables the construction of a synthetic dataset consisting of UAV images and corresponding segmentation masks for the task of flood segmentation. To achieve this, we employed two distinct generative strategies, subsequently integrating their outputs into a unified dataset. In both strategies, we ensured that the generated images comprehensively represent the range of flood scenarios observed in real-world datasets (see Figure 1). This was achieved by systematically varying key semantic attributes including scene context (urban/peri-urban vs. rural), camera viewpoint (presence or absence of sky), and flood severity (from light inundation to extensive submersion).
The first approach leverages a text-to-image generative model, enabling the synthesis of UAV imagery that encapsulates specific flood-related characteristics, explicitly defined through textual descriptions. The second approach employs an image-to-image translation paradigm, wherein segmentation masks derived from real data serve as structural constraints for synthesizing realistic UAV images via an inpainting method. As a final step in the synthetic dataset creation pipeline, in order to ensure the fidelity and realism of the generated synthetic images, we apply a filtering process, based on outlier detection mechanisms, that aims to discard unrealistic synthetic images. An overview of the synthetic flood dataset generation method is shown in Figure 3.
For both generative strategies, we selected Stable Diffusion [47] over other synthetic data generation methods. Although VAEs and GANs are also common choices for synthetic data generation, Stable Diffusion presents advantages in image quality, training stability, and controllability across diverse domains [48,56]. It enables fine-grained semantic control via prompt conditioning, supports diverse output generation through simple stochastic sampling, and natively handles both text-to-image and image inpainting tasks. These capabilities allowed us to synthesize realistic and varied flood scenarios with high fidelity, which are difficult to achieve with VAEs and GANs due to their limited controllability and domain consistency [46,57]. Thus, Stable Diffusion was a more suitable choice for generating synthetic data to complement real-world flood datasets.

3.1. Text-to-Image Synthesis Component

In the first direction, given a text-to-image synthesis models, noted as G s , which in our case was the Stable Diffusion [47], and a prompt P i from a predefined set P , the output of the process is a generated synthetic image I s :
I s = G s ( P i ) , P i P .
Stable Diffusion [47] is a generative model that synthesizes high-quality images by reversing a gradual noising process through a series of denoising steps. Initially, an image is represented as pure noise, and the model iteratively refines it over a fixed number of steps, by progressively removing noise while conditioning on a text prompt or other guidance (e.g., semantic maps or class labels). At each step, the model predicts the noise component and subtracts it from the current image estimate, gradually revealing a coherent image that aligns with the input conditions. We have found that 50 steps are enough to generate a visually realistic image that reflects the intended structure and content, in our case, a flooded environment, generated from an entirely random starting point, as depicted in Figure 4.
The quality of the generated outputs and the ability of the synthetic image dataset to adequately capture the variation inherent in UAV images depicting flooded areas are heavily influenced by the chosen input prompts. To ensure that the dataset encompasses a diverse range of scenarios, we systematically varied the prompts along three key attributes, while controlling flood severity with specific word choices (partially, fully, some affected by flooding, etc.):
  • Rural vs. urban/peri-urban environment: The images were conditioned to represent either a rural or urban/peri-urban landscape.
  • Sky presence: The synthesized images either included or excluded visible sky regions.
  • Flooded and non-flooded buildings: We controlled the number of flooded buildings in the generated images to ensure a diverse range of flooding scenarios.
The following are examples of prompts used for data generation:
  • “Aerial view of a flooded urban area with high-rise buildings and streets underwater."
  • “Drone footage of a rural landscape with scattered houses, some affected by flooding, others dry.”
  • “Low-altitude remotely sensed image depicting an urban neighborhood with partially submerged homes and roads.”
  • “UAV view of a countryside area with a river overflow flooding nearby fields and farmhouses.”
Since the generated outputs lacked corresponding ground truth segmentation masks, we employed an unsupervised mask generation method to derive pseudo-labels (PL). Specifically, to generate the unsupervised pseudo-labels, we relied on the UFS-HT-REM method, as described in [18]. This method autonomously identifies flooded and non-flooded regions based on color and texture properties in aerial imagery. A hysteresis threshold T ( · ) is applied on a probabilistic map P M ( · ) derived from color, intensity, and reflectance information C I R ( · ) :
P L s = T ( P M ( C I R ( I s ) ) ) .
where P L s is the resulting pseudo-label binary mask of the synthetic image I s , with each pixel value defined as
P L s ( x , y ) = 1 , if ( x , y ) belongs to a flooded region 0 , otherwise .
This process ensures robust segmentation by leveraging statistical color distributions and intensity variations to separate flood water from background, even in visually complex images. The final output of the process is a synthetic UAV image I s and its corresponding binary pseudo-label mask P L s .

3.2. Image Inpainting from Segmentation Masks Component

An alternative approach to generating synthetic UAV images was image inpainting using real segmentation masks. This method involved using segmentation masks from a real dataset and applying an inpainting algorithm to synthesize realistic UAV imagery corresponding to the given segmentation structure. Using real flood segmentation masks as priors provides a more grounded structural context, ensuring that the generated images reflect plausible flood extents and spatial layouts as observed in real-world scenarios. This leads to synthetic data that are more representative and structurally aligned with actual flood patterns, thereby improving its utility for training segmentation models.
Overall, the process can be formulated as follows. Given a flood-related image, I r , selected from a real-world UAV flood segmentation dataset, and its corresponding ground truth segmentation mask G T , employ an image inpainting model G i p to generate a semi-synthetic UAV image I i p with the real flood area untouched, thus inpainting the labeled background, conditioned on a textual context prompt, P j :
I i p = G i p ( G T , P j ) .
Following the previous approach, we again adopted Stable Diffusion [47] in our workflow as an image inpainting model in a text-guided image-to-image translation framework. The prompts were selected based on the image category, ensuring consistency with the expected scene semantics. Prompt context was similar to the previous process, to ensure a thematic consistency between both synthetic image generation schemes.
Since the generated image I i p does not always exhibit a perfect one-to-one correspondence with the input ground truth G T , we apply the unsupervised pseudo-labeling method once again to derive a refined segmentation mask P L i p for the newly synthesized image:
P L i p = T ( P M ( C I R ( I i p ) ) ) .
We observed that this process better ensures that the final segmentation pseudo-labels P L i p more accurately reflect the flooded and non-flooded regions present in the semi-synthetic, inpainted images.

3.3. Post-Generation Filtering for Enhanced Dataset Fidelity

Following the generation of two synthetic image sets
S D s = { ( I s , P L s ) | I s = G s ( P i ) , P L s = T ( P M ( C I R ( I s ) ) ) } ,
and
S D i p = { ( I i p , P L j ) | I i p = G i p ( G T , P j ) , P L i p = T ( P M ( C I R ( I i p ) ) ) } ,
we merged both sets into a unified synthetic dataset:
S D a l l = S D s S D i p .
To improve the unified synthetic dataset fidelity, we applied the FINCH [58,59] clustering algorithm to S D a l l , retaining only images from the first partition and discarding those that formed singleton clusters. FINCH is an ideal choice since it is a parameter-free hierarchical clustering algorithm that constructs clusters in a cluster-number agnostic way using a simple yet effective strategy based on first-neighbor relations. Specifically, each data point is initially linked to its nearest neighbor, and transitive links are formed such that if point A links to B and B links to C, all three belong to the same cluster. This process yields a partitioning of the data without requiring the number of clusters or other hyperparameters to be specified. The procedure is applied recursively to the resulting centroids, producing a hierarchy of clusterings that capture increasingly abstract groupings.
This choice ensures that filtering is applied early in the generation pipeline, where redundancy and noise are more prevalent, thereby preventing low-quality or outlier samples from propagating into later stages. Next, we encoded each synthetic image into a feature vector of size 1 × 1024 using a pre-trained ResNet50 model.
To further refine the dataset, we applied FINCH clustering to a real-world dataset and extracted cluster centers. We then computed pairwise Euclidean distances between the real cluster centers and the synthetic image feature vectors. Synthetic images were assigned to the closest cluster, and those failing to align with any real-world cluster were discarded:
d ( I s y n , C r e a l ) = min c | | f ( I s y n ) c 2 , c C r e a l
where f ( I s y n ) represents the feature embedding of a synthetic image and C r e a l is the set of real-world cluster centers.
After applying the filtering process, where images failing to align with real-world cluster structures are discarded, the final filtered synthetic dataset S D f i l t is defined as
S D f i l t = { ( I , P L ) S D a l l d ( f ( I ) , C r e a l ) τ } ,
where I is a synthetic or semi-synthetic image, and P L is its corresponding pseudo-label segmentation mask, f ( I ) represents the feature embedding of the image I, extracted using a pre-trained ResNet50, C r e a l is the set of cluster centers obtained from the real-world dataset using the FINCH algorithm, d ( · , · ) denotes the Euclidean distance between the synthetic image embedding and the closest real-world cluster center, and τ is a distance threshold that determines the inclusion of synthetic images based on their proximity to the real-world data. The resulting set S D f i l t , has a size of 557 images (279 from SDs, 278 from SDip) and their corresponding pseudo-label masks, whereas pre-filtering size was 580 images (290 from SDs, 290 from SDip) and their corresponding pseudo-label masks. Synthetic image sets are publicly available (https://sites.google.com/site/costaspanagiotakis/research/flood-detection, accessed on 3 June 2025).
In Figure 5, representative examples of synthetic (a) and semi-synthetic (e) images, and their corresponding pseudo-label segmentation masks overlaid in blue, (b) and (f), are shown. For the inpainting process of the semi-synthetic images, ground truths (d) were used, derived from real-world images (c). The synthetic images exhibit a high degree of diversity in terms of environmental conditions, flood severity, and background complexity, as well as variation in camera angles and perspectives. This diversity is essential to enhance the model’s robustness and generalizability across a wide range of flooding scenarios. Moreover, the synthetic data were generated with particular attention to visual realism and structural similarity to real-world flood imagery, enabling the more effective transfer of knowledge when such images are used in training. As previously mentioned, when inpainting is applied, the original ground truth annotations become invalid, as the generative process may not only modify the background but also extend or alter the spatial extent of the flooded regions, thereby introducing discrepancies between the modified image and its initial label.
Figure 6 presents examples of synthetic (a) and semi-synthetic (d) images which did not pass the filtering process. Also, the real-world images (b) and their respective ground truths (c), from which the semi-synthetic versions were generated, are also provided for reference. The discarded synthetic samples exhibit significant deviations from realistic flood scenes, including structural inconsistencies, the unnatural repetition of patterns, excessive blurriness, or visual characteristics resembling stylized artwork (e.g., oil paintings). These artifacts resulted in the images being classified as unrealistic according to our established outlier scoring scheme, thereby justifying their removal from the training set.
Outlier threshold, τ : we employ a z-score-based filtering mechanism grounded in the distribution of distances between synthetic features and real cluster centers. Specifically, for each synthetic feature vector, we computed its minimum Euclidean distance to the closest cluster center derived from FINCH clustering on real data. Let μ and σ denote the mean and standard deviation of these minimum distances, respectively. A synthetic sample i was retained if its distance d i satisfied:
d i μ + k · σ ,
where k is a hyperparameter controlling the strictness of the filtering (we set k = 1.5 in our experiments). This statistical approach adapts the filtering threshold to the inherent distribution of the distances, avoiding arbitrary cutoffs and promoting a more robust selection of semantically similar synthetic samples.

4. Experimental Setup

To assess the generalization capability of a segmentation model trained on synthetic flood imagery, we conducted a series of controlled experiments using state-of-the-art deep learning architectures. Specifically, we selected three top-performing segmentation models which were trained under six different data configurations: (a) exclusively on the real-world dataset with the actual ground truth D r = ( I r , G T ) , (b) on the real-world dataset with pseudo-labels assuming no ground truth is available D p l = ( I r , P L r ) , (c) on the synthetic dataset S D s (see Equation (6)) with corresponding pseudo-labels ( P L s ), (d) on the semi-synthetic dataset S D i p (see Equation (7)) with corresponding pseudo-labels ( P L i p ), (e) on the combination of real-world and both synthetic and semi-synthetic datasets D r S D a l l ( S D a l l as defined in Equation (8)) with the actual ground truth and corresponding pseudo-labels ( P L a l l ), and (f) on the combined real-world and filtered synthetic and semi-synthetic datasets D r S D f i l t ( S D f i l t as defined in Equation (10)) with the actual ground truth and corresponding pseudo-labels ( P L f i l t ) . The evaluation was performed on an independent real-world dataset with respective expert annotations, which was not used during training, to measure the effectiveness of the synthetic data in bridging the domain gap.

4.1. Datasets and Methods

Datasets: For the real-world dataset baselines, we employed two publicly available datasets that depict flood-affected regions, each accompanied by ground truth segmentation masks delineating flooded areas. These datasets consist of aerial imagery captured by UAVs and helicopters, encompassing a diverse range of environmental contexts, including urban, peri-urban, and rural landscapes. The images exhibit significant variability in scene composition, featuring elements such as vegetation, rivers, buildings, roads, mountainous terrain, and the sky. Additionally, they were acquired from multiple altitudes and viewing angles, ensuring a comprehensive representation of flood scenarios. Notably, both datasets maintain a similar balance between flood and background pixels, which mitigates potential class imbalance issues during model training.
As a baseline training set D r , we opted to use the well-known Flood Area Dataset (FAD) [11]. This dataset comprises 290 RGB images, accompanied with manually annotated segmentation masks. The dataset exhibits variability in image resolution and dimensions. Contributing to the dataset’s heterogeneity is its diverse range of environmental contexts, including 203 urban and peri-urban scenes and 87 rural scenes. In terms of visual composition, 108 images contain full or partial views of the sky, while 182 images lack any visible sky. This diversity supports robust model training across varying landscape types and viewing conditions.
To assess the generalization capability of the segmentation models, we utilized only for inference the Flood Semantic Segmentation Dataset (FSSD) [13] as an independent test set, which consists of 663 RGB images and corresponding ground truth segmentation masks. Similarly to the FAD, this dataset comprises images obtained from UAVs portraying diverse flooded scenes captured from various camera perspectives. The image sizes and resolutions also vary, but were all resized and, if necessary, zero-padded, to 512 × 512 by the dataset creator. All 663 images were used as our test set. Representative samples from both datasets, along with the corresponding ground truth masks, are illustrated in Figure 7.
Flood segmentation methods: We trained three well-established Convolutional Neural Networks (CNNs) for semantic segmentation tasks, and two transformer-based models.
We started with the following CNN architectures: DeepLabV3, FCN-ResNet50, and U-Net. DeepLabV3 [60] incorporates atrous (dilated) convolution and Atrous Spatial Pyramid Pooling (ASPP) to capture multi-scale contextual information without reducing spatial resolution. Built on a ResNet backbone, it is effective in segmenting complex scenes and fine structures. FCN-ResNet50 follows the Fully Convolutional Network design [61] enhanced with a ResNet-50 backbone. This architecture performs dense prediction by replacing fully connected layers with convolutions, while residual and skip connections help preserve spatial detail. U-Net, introduced in [62], features a symmetric encoder–decoder structure with skip connections between corresponding layers. This design enables precise localization and efficient learning from limited data, making it especially suitable for segmentation tasks.
Furthermore, we included two Transformer-based models in our evaluation: SegFormer-B0 and Swin-Tiny with UPerNet (Swin-T). SegFormer-B0 [63] combines a lightweight MixVision Transformer encoder with a UPerNet-style decoder. The encoder avoids positional embeddings for improved generalization and employs a hierarchical design that efficiently captures multi-scale context. The B0 variant offers a strong performance–efficiency trade-off, making it suitable for training from scratch with limited data while maintaining good segmentation quality. Swin-T leverages the Swin Transformer’s hierarchical architecture with shifted window attention, enabling the effective modeling of both local and global dependencies [64]. Coupled with the UPerNet decoder, which aggregates multi-scale features, this model delivers a strong performance in dense prediction tasks and excels at delineating sharp object boundaries, but at a greater computational cost.

4.2. Training Protocol

Implementation Details: From the range of diffusion models, we selected Stable Diffusion due to its superior ability to generate photorealistic aerial and environmental flood scenes that closely resemble UAV imagery, including urban and rural landscapes with varied perspectives (e.g., top–down, oblique). Unlike models such as Midjourney or DALL·E, Stable Diffusion offers fine-grained prompt control, open source accessibility, fine-tuning capabilities, native support for image inpainting enabling targeted scene manipulation and seamless integration into PyTorch-based pipelines.
In this study, we employed Stable Diffusion v1.5 [47], chosen for its balance between visual fidelity, controllability, and computational efficiency. Trained on the extensive and diverse LAION-5B dataset [65], this version is particularly effective at generating realistic and coherent environments, accurately capturing natural elements such as water, rain, clouds, and terrain, making it well suited for synthesizing UAV-like flood imagery. Compared to earlier versions, it offers improved stability, reduced artifacts, and greater prompt interpretability, supporting prompts up to 77 tokens, which enabled us to steer the generation process toward specific flood scenarios (e.g., “urban flooding” or “flooded rural roads with debris”). These characteristics ensure high-quality synthetic data with realistic textures and geometries, which is critical for effective training of segmentation models. Additionally, the relatively compact architecture of the model (860 M UNet, 123 M text encoder) allows for efficient deployment on consumer-grade GPUs.
All models were implemented in PyTorch (v. 2.6.0, CUDA 12.6) and trained from scratch for 100 epochs with a random generated batch of eight images, using an initial learning rate of 1 × 10 3 . Each utilized dataset was randomly partitioned into 90% for training and 10% for validation. The Dice loss function was employed as the objective to address class imbalance, and model performance was evaluated using the accuracy on the validation set.
For the CNN models, a dynamic learning rate adjustment strategy was employed, whereby the learning rate was reduced by a factor of 0.5, if no improvement in validation performance was observed over five consecutive epochs. Optimization was conducted using the Adam optimizer with a weight decay of 1 × 10 4 to prevent overfitting by penalizing large weights and encouraging model generalization. Adam was selected due to its adaptive learning rate strategy, which computes individual learning rates for each parameter based on the estimates of first and second moments of the gradients. This facilitates efficient convergence and robustness in training deep neural networks, particularly in high-dimensional and non-convex optimization landscapes. The model weights were initialized using the Kaiming normal initialization, which is specifically designed for layers with ReLU activations. This initialization strategy maintains the variance of activations across layers, thereby promoting stable gradient flow during training and preventing issues such as vanishing or exploding gradients.
For the transformer-based models, we followed the suggested training configurations of the original papers [63,64], and utilized AdamW as the optimizer with a weight decay of 1 × 10 2 , and a cosine learning rate scheduler with warm-up. AdamW is a variant of Adam that decouples weight decay from the gradient update. This modification improves generalization and is particularly effective for training transformer architectures. Like Adam, it adapts learning rates based on the first and second moments of the gradients, enabling stable and efficient optimization in complex, high-dimensional settings. The cosine learning rate scheduler with warm-up gradually increases the learning rate during an initial warm-up phase, followed by a cosine decay, which helps prevent early training instability and encourages smoother convergence. The model weights were initialized using the standard procedures provided by the respective implementations, without any pre-training.
Notably, improved model convergence and higher validation accuracy were observed when training was performed on a combined dataset comprising real-world, synthetic, and semi-synthetic images (see also Section 5). In this setting, each training batch was composed of a fixed ratio of samples—specifically, two randomly selected images from the real-world dataset and three images each from the synthetic and semi-synthetic subsets. This sampling strategy enabled the model to benefit from the diversity and volume of synthetic data while maintaining grounding in real-world examples, effectively enhancing generalization and training stability.
Training was performed on a system equipped with an Intel i7 CPU (2.3 GHz), 40 GB of RAM, and two NVIDIA Quadro RTX 4000 GPUs. The total training times ranged from less than an hour to approximately four hours, depending on the deep learning architecture and the size of the training dataset.
Dataset pre-processing: In the case of model training in real-word data, as previously stated, we utilized the FAD dataset. In the absence of ground truth, corresponding pseudo-label segmentation masks were automatically generated using an adapted UFS-HT-REM [18] pipeline. All images were resized to 416 × 320 pixels to speed up training, and normalized to zero mean and unit variance. These processes were also applied on the synthetic dataset, as well as on the FFSD dataset that is used for model performance evaluation.
Data augmentation: To mitigate the limited training data, we applied on-the-fly augmentation with a per-transformation probability of 0.5. Augmentations included horizontal flipping, random rotations within [ 15 , 15 ] degrees, and additive uniform noise in the range [ 0.2 , 0.2 ] . Over 100 epochs, this strategy effectively expanded training variability.

5. Experimental Results

In this section, we present an evaluation of the proposed synthetic dataset generation method in the UAV flood segmentation task. First, we assess the effectiveness of training segmentation models using only synthetic data, comparing their performance against the same models trained on real-world datasets. In the second part of this section, we analyze the role of prompt structure and semantic context in the quality and utility of the generated synthetic images. Additionally, we evaluate the similarity between the synthetic and real-image feature distributions by providing clustering statistics.

5.1. Impact of Synthetic Data on Model Performance

We evaluated the impact of the proposed synthetic dataset creation framework on the performance of three widely used convolution-based semantic segmentation architectures: DeepLabV3 [60], FCN-ResNet50 [61], and U-Net [62], and on two top-performing transformer-based architectures, namely SegFormer-B0 [63] and Swin-Tiny + UPerNet (Swin-T) [64]. For this, each model was trained exclusively on both intermediate synthetic datasets (text-to-image, S D s , and image inpainting, S D i p ), on the combined real-world and synthetic dataset, before ( D r S D a l l ) and after the filtering process ( D r S D f i l t ), and was compared against its counterpart trained on the real-world dataset with the actual ground truth D r , and with pseudo-label segmentation masks D p l instead of the ground truth. As reported in Table 1, Table 2 and Table 3, across all convolutional architectures, we observed that models trained solely on synthetic data achieved segmentation performance remarkably close to those trained on real-world data, and in case of the U-Net even better. On average, the observed performance was above 70 % in F1-score, with a drop in the range of 2–5% compared to training with real-world data, and a raise of 1.07% in case of the U-Net, highlighting the high fidelity and generalizability of the generated synthetic samples.
A similar performance impact is observed for the transformer-based architectures, as shown in Table 4 and Table 5, where on average performance was also above 71 % in F1-score. In these cases, a model trained solely on synthetic data, resulted in an average performance difference of 4 % compared to a version trained on real-world data (rows 3 and 5–6 in Table 4 and Table 5). Interestingly, while the general performance trend between real and synthetic datasets holds consistently across both convolutional and transformer-based models, the absolute segmentation performance of the transformer-based architectures was found to be comparable to or slightly lower than that of the convolutional models. This can be partly attributed to the fact that all models were trained under identical learning configurations, which might have not allowed for the transformer-based methods to reach their full potential, given the fact that recent findings suggest that transformer-based architectures typically require longer training schedules to fully converge, due to their higher capacity and weaker inductive biases. Moreover, transformers tend to benefit more from large, diverse datasets, because of their global attention mechanisms, whereas convolutional models can generalize better under limited data due to their strong local spatial priors.
Table 1. DeepLabV3 performance on the test dataset (FSSD) based on training with the real dataset and ground truths ( D r ), the real dataset and pseudo-labels ( D p l ), the text-to-image synthetic dataset ( S D s ), the semi-synthetic image inpainting dataset ( S D i p ), and their union, unfiltered ( S D a l l ) and filtered ( S D f i l t ), when it is combined with the real-world dataset D r . The best-performing values for each metric are highlighted in bold.
Table 1. DeepLabV3 performance on the test dataset (FSSD) based on training with the real dataset and ground truths ( D r ), the real dataset and pseudo-labels ( D p l ), the text-to-image synthetic dataset ( S D s ), the semi-synthetic image inpainting dataset ( S D i p ), and their union, unfiltered ( S D a l l ) and filtered ( S D f i l t ), when it is combined with the real-world dataset D r . The best-performing values for each metric are highlighted in bold.
Training Test Metrics
DatasetAcc (%)IoU (%)Pr (%)Rec (%)F1 (%)
Dr79.4860.0367.3984.6275.03
Dpl78.0358.7565.0485.8574.01
SDs76.4855.0264.4778.9770.99
SDip74.5454.7060.8584.4270.72
Dr ∪ SDall79.2160.9465.8989.0275.73
Dr ∪ SDfilt79.5561.3466.3489.0676.04
Among the evaluated architectures, we observed that DeepLabV3 exhibited the highest performance drop when trained solely on synthetic data, while U-Net consistently achieved a slight increase in performance—approximately 1% across all evaluation metrics—compared to its training on real-world data. This divergence in behavior can be attributed to architectural differences: DeepLabV3, with its atrous spatial pyramid pooling and deeper encoder, may rely more heavily on fine-grained real-world features and textures that the synthetic data only partially captures. In contrast, U-Net’s symmetric encoder–decoder structure with skip connections may better exploit the spatial coherence and regularity present in the synthetic masks and images, thus benefiting from the structured nature of the generated dataset.
Table 2. FCN-ResNet50 performance on the test dataset (FSSD) based on training with the real dataset and ground truths ( D r ), the real dataset and pseudo-labels ( D p l ), the text-to-image synthetic dataset ( S D s ), the semi-synthetic image inpainting dataset ( S D i p ), and their union, unfiltered ( S D a l l ) and filtered ( S D f i l t ), when it is combined with the real-world dataset D r . The best-performing values for each metric are highlighted in bold.
Table 2. FCN-ResNet50 performance on the test dataset (FSSD) based on training with the real dataset and ground truths ( D r ), the real dataset and pseudo-labels ( D p l ), the text-to-image synthetic dataset ( S D s ), the semi-synthetic image inpainting dataset ( S D i p ), and their union, unfiltered ( S D a l l ) and filtered ( S D f i l t ), when it is combined with the real-world dataset D r . The best-performing values for each metric are highlighted in bold.
TrainingTest Metrics
DatasetAcc (%)IoU (%)Pr (%)Rec (%)F1 (%)
Dr83.6066.9071.6590.9880.16
Dpl83.6467.6770.7294.0080.72
SDs81.3564.8067.4894.2278.64
SDip81.2864.4467.6593.1478.38
Dr ∪ SDall83.7467.8270.8494.0980.83
Dr ∪ SDfilt84.2268.1272.0892.5481.04
Table 3. U-Net performance on the test dataset (FSSD) based on training with the real dataset and ground truths ( D r ), the real dataset and pseudo-labels ( D p l ), the text-to-image synthetic dataset ( S D s ), the semi-synthetic image inpainting dataset ( S D i p ), and their union, unfiltered ( S D a l l ) and filtered ( S D f i l t ), when it is combined with the real-world dataset D r . The best-performing values for each metric are highlighted in bold.
Table 3. U-Net performance on the test dataset (FSSD) based on training with the real dataset and ground truths ( D r ), the real dataset and pseudo-labels ( D p l ), the text-to-image synthetic dataset ( S D s ), the semi-synthetic image inpainting dataset ( S D i p ), and their union, unfiltered ( S D a l l ) and filtered ( S D f i l t ), when it is combined with the real-world dataset D r . The best-performing values for each metric are highlighted in bold.
TrainingTest Metrics
DatasetAcc (%)IoU (%)Pr (%)Rec (%)F1 (%)
Dr81.0362.6268.9587.2177.01
Dpl80.3363.0466.6592.0777.33
SDs81.4564.0468.5690.6878.08
SDip81.1164.0467.6392.3478.08
Dr ∪ SDall81.9165.5268.2194.3279.17
Dr ∪ SDfilt86.9772.0076.8591.9483.72
Table 4. SegFormer-B0 performance on the test dataset (FSSD) based on training with the real dataset and ground truths ( D r ), the real dataset and pseudo-labels ( D p l ), the text-to-image synthetic dataset ( S D s ), the semi-synthetic image inpainting dataset ( S D i p ), and their union, unfiltered ( S D a l l ), and filtered ( S D f i l t ), when it is combined with the real-world dataset D r . The best-performing values for each metric are highlighted in bold.
Table 4. SegFormer-B0 performance on the test dataset (FSSD) based on training with the real dataset and ground truths ( D r ), the real dataset and pseudo-labels ( D p l ), the text-to-image synthetic dataset ( S D s ), the semi-synthetic image inpainting dataset ( S D i p ), and their union, unfiltered ( S D a l l ), and filtered ( S D f i l t ), when it is combined with the real-world dataset D r . The best-performing values for each metric are highlighted in bold.
TrainingTest Metrics
DatasetAcc (%)IoU (%)Pr (%)Rec (%)F1 (%)
Dr78.4061.4763.7094.6176.14
Dpl79.4362.4365.1093.8476.87
SDs79.0361.1265.3090.5275.87
SDip74.3457.6359.1795.6873.12
Dr ∪ SDall80.1363.6165.6695.3377.76
Dr ∪ SDfilt81.2264.9666.9995.4878.74
A particularly noteworthy outcome of our experimental analysis is the consistent performance improvement observed across all examined segmentation models, when trained on the combined dataset comprising both real-world data (images and manually annotated masks) and synthetic data (generated images and pseudo-label masks). These gains range between 1% and 6% in the F1-score for the convolutional architectures and between 3% and 7% for the transformer-based architectures. This observed behavior underscores the complementary nature of the synthetic dataset. The inclusion of synthetically generated samples, which capture diverse flood-affected scenarios and augment underrepresented visual patterns, likely aids in regularizing the models and enhancing their generalization capabilities. Additionally, the pseudo-label masks, though automatically generated, provide reasonably accurate supervision signals that help the models learn more robust decision boundaries. This result highlights the value of synthetic data as a scalable and effective means to enrich limited annotated datasets in domain-specific applications such as disaster response.
Table 5. Swin-T performance on the test dataset (FSSD) based on training with the real dataset and ground truths ( D r ), the real dataset and pseudo-labels ( D p l ), the text-to-image synthetic dataset ( S D s ), the semi-synthetic image inpainting dataset ( S D i p ), and their union, unfiltered ( S D a l l ) and filtered ( S D f i l t ), when it is combined with the real-world dataset D r . The best-performing values for each metric are highlighted in bold.
Table 5. Swin-T performance on the test dataset (FSSD) based on training with the real dataset and ground truths ( D r ), the real dataset and pseudo-labels ( D p l ), the text-to-image synthetic dataset ( S D s ), the semi-synthetic image inpainting dataset ( S D i p ), and their union, unfiltered ( S D a l l ) and filtered ( S D f i l t ), when it is combined with the real-world dataset D r . The best-performing values for each metric are highlighted in bold.
TrainingTest Metrics
DatasetAcc (%)IoU (%)Pr (%)Rec (%)F1 (%)
Dr81.8564.9368.6992.2278.74
Dpl81.2461.6670.7282.8076.28
SDs78.6756.9468.2777.4372.57
SDip77.1155.9865.1679.9071.78
Dr ∪ SDall83.5366.3972.1189.3279.80
Dr ∪ SDfilt84.8868.6973.6691.0681.44
The introduction of our filtering strategy following the initial creation of the synthetic dataset (row 5 vs. row 6 in Table 1, Table 2 and Table 3) further enhances model performance across all segmentation architectures. Specifically, we observe improvements ranging from 1% to 4% in key evaluation metrics when the filtered synthetic dataset is used instead of the unfiltered one. Among the evaluated models, U-Net consistently exhibits the highest performance gain, suggesting that its architecture may be particularly sensitive to noisy or low-quality training examples. By eliminating synthetic samples with low semantic alignment to real data distributions, the filtering stage effectively increases the signal-to-noise ratio in the training set, leading to better convergence and generalization. This demonstrates the importance of curating synthetic data not only in terms of diversity, but also in maintaining fidelity to real-world distributions.
Among the evaluated convolutional architectures, as observed in Figure 8, DeepLabV3 yielded the lowest performance when trained solely on synthetic data, though it was able to marginally surpass training on real-world data when a combined real and synthetic dataset was used. FCN-ResNet50 consistently achieved the highest performance in most scenarios, yet its performance declined slightly under synthetic-only training conditions. U-Net, in contrast, demonstrated notably stronger performance with synthetic data and significantly outperformed the other deep learning models when trained with a filtered combination of real and synthetic data. This can be attributed to U-Net’s encoder–decoder structure with skip connections, which excels at capturing both fine-grained local details and global spatial context. Such architectural characteristics are particularly well-suited for segmentation tasks involving structured patterns. Consequently, the U-Net architecture excels in our synthetic flood imagery that features regular shapes and consistent textural cues, as generated via Stable Diffusion. U-Net’s inductive biases align well with the spatial regularity present in the synthetic data, enhancing its ability to generalize effectively in this context. Statistically, for this model family, comparing the F1 scores, we have obtained 58.22% over all images better inferences with U-Net, 29.26% with FCN-ResNet50, and only 12.67% with DeepLabV3 indicating it is the worst performing model when trained with combined real and synthetic data ( D r S D f i l t ). Note that inferences can excel in more than one model.
Regarding the transformer-based methods, as shown in Figure 8 and in Table 4 and Table 5, SegFormer-B0 exhibited constant high performance when trained on the synthetic images, comparable to its performance when trained solely on real data. This can be attributed to its lightweight and efficient design, which integrates a simplified hierarchical structure and a minimal decoder, allowing it to generalize well even under limited data conditions and shorter training schedules. On the other hand, Swin-T demonstrated the lowest performance when trained only on synthetic data, even amongst all examined models, potentially attributed to its more complex architecture that relies on shifted window attention and multi-stage hierarchical processing, which, while powerful, has been shown to require larger datasets, longer training schedules, and careful fine-tuning to reach its full potential. This limitation is further exacerbated by the inclusion of the heavy decoder head (UPerNet), which adds additional optimization challenges and increases sensitivity to domain shifts. These characteristics combined with the wide representation span of the images present in the synthetic dataset, as shown in Figure 9, and the relatively small size of the datasets, can justify its lower performance. Conclusively, among the transformer-based models, comparing the F1 scores, we have obtained 66.37% over all images better inferences with Swin-T, and 33.63% with SegFormer-B0, when trained with combined real and synthetic data ( D r S D f i l t ).
Overall, the vast majority of the examined models across both architectural categories exhibit comparable or even superior performance when trained solely on synthetic data compared to when trained exclusively on real data. Furthermore, we observe that all models achieve substantial performance improvements when trained on the combined real and synthetic datasets. These two observations underscore the valuable contribution of synthetic data incorporation in the learning process and highlight the effectiveness and practical utility of the proposed synthetic data generation framework in enhancing model generalization and segmentation performance. Taken together, based on F1 score comparisons across all images, U-Net achieved superior segmentation performance in 38% of the cases, followed by Swin-T (23.98%), FCN-ResNet50 (18.1%), SegFormer-B0 (12.97%), and DeepLabV3 (6.94%). These results highlight U-Net as the best-performing model when trained on the combined real and filtered synthetic dataset ( D r S D f i l t ), as also illustrated in Figure 8.
Evaluation of synthetic data impact under controlled settings: In order to further highlight the contribution of synthetic data in the learning process, we conducted an additional experiment using U-Net as the examined model, since, based on the reported results, it is the one that exhibits the largest performance gains in general. In this new experiment, we fixed the total number of training images to match the real-only condition. Specifically, we created a new training configuration composed of 145 real images and 145 synthetic images, totaling 290 samples, the same as in the real-only case ( D r in row 3 of Table 3). All training hyperparameters and iteration counts were kept constant across configurations. The results of this experiment, depicted in Table 6, show that this configuration achieves a better performance ( 1.3 % increase in F1-score) than the real-only case, indicating that synthetic data can partially replace real data without loss of segmentation accuracy. This provides further evidence that the gains observed in the configuration combining real and synthetic data (last row of Table 3) are not merely due to increased data volume but rather reflect the complementary utility of the synthetic images.
Table 6. U-Net performance on the test dataset (FSSD) based on training with randomly selected images 50%, 25%, 25% from the real, synthetic, and semi-synthetic dataset (50% Dr ∪ 25% SDs ∪ 25% SDip) in comparison to training with the real dataset only ( D r ). Both training datasets are of the same size. The best-performing values for each metric are highlighted in bold.
Table 6. U-Net performance on the test dataset (FSSD) based on training with randomly selected images 50%, 25%, 25% from the real, synthetic, and semi-synthetic dataset (50% Dr ∪ 25% SDs ∪ 25% SDip) in comparison to training with the real dataset only ( D r ). Both training datasets are of the same size. The best-performing values for each metric are highlighted in bold.
TrainingTest Metrics
Dataset (290 Images)Acc (%)IoU (%)Pr (%)Rec (%)F1 (%)
Dr81.0362.6268.9587.2177.01
50% Dr ∪ 25% SDs ∪ 25% SDip81.2864.3067.8292.5378.27

5.2. Real and Synthetic Dataset Similarity, Role of Prompt Semantics in Dataset Quality

Real and synthetic data similarity: To quantitatively assess the distributional similarity between real and synthetic datasets, we computed the Maximum Mean Discrepancy (MMD) between their respective feature representations. Specifically, we extracted deep features using a pre-trained ResNet50 model and computed the MMD between the real dataset and two variants of the synthetic dataset. The resulting scores were MMD D r , SD s = 0.0545 and MMD D r , SD i p = 0.0347 . Both scores indicate a reasonable alignment with the real data distribution, as MMD values below 0.1 are typically indicative of good distributional similarity in high-dimensional spaces. Notably, the second synthetic dataset demonstrates significantly closer alignment to the real data, suggesting improvements in generation fidelity—possibly due to enhanced prompt structure, better context grounding, or more effective filtering. These results support the hypothesis that high-quality synthetic data, when properly curated, can closely mimic real-world distributions and serve as a valuable resource for training deep segmentation models.
Additionally, as a second way to assess the similarity between a real-world dataset (Dr) and two synthetic datasets (SDs and SDip) with a more direct and visual manner, we also used the feature embeddings from each image and applied principal component analysis (PCA) for dimensionality reduction and visualization. The resulting 2D PCA plot, shown in Figure 9, demonstrates a substantial overlap between the three datasets in the feature space (best-fit ellipses overlap), indicating a high degree of visual similarity. Notably, the synthetic datasets exhibit significant alignment with the real-world distribution, suggesting that the synthetic data generation processes effectively capture the global visual characteristics of real imagery. Among them, SDip shows a tighter overlap with Dr, which can be attributed to the ground truth mask prior, while SDs exhibits a slightly broader spread, which may imply higher visual diversity or variability in synthesis quality. This kind of deviation may reflect that, while the synthetic dataset captures much of the variance of the real data (hence the overlap), it might also be exploring new areas or patterns in the feature space that are not fully represented in the real data. This is also suggested by the slight deviation in the orientation of the fitted ellipses between the real and synthetic datasets. The novel pattern assumption appears to be verified by the model performance increase when training on the union of the real and synthetic datasets. These findings support the potential utility of synthetic datasets as substitutes or complements to real-world data for training DL models, especially in scenarios where labeled real data is limited or costly to obtain.
Prompt structure and semantics in synthetic image quality: The structure and semantics of textual prompts play a critical role in guiding the fidelity of generated synthetic images. As shown in Figure 10, minor variations in phrasing, specificity, or contextual richness can substantially affect visual realism and alignment with the intended flood scenario. Detailed prompts that include spatial relationships, lighting conditions, environmental context (e.g., urban vs. rural), and object-level cues (e.g., “partially submerged tractors”, “muddy floodwater”, “sky reflections”) consistently result in higher-quality images. Conversely, vague or semantically sparse prompts often lead to artifacts, inconsistencies in water boundaries, or unnatural object placements (e.g., “flooded city” or “flooded river”).
For instance, as illustrated in Figure 10, the use of a vague or underspecified prompt in the text-to-image synthesis pipeline leads to the generation of semantically and structurally inconsistent elements, such as the appearance of building rooftops that were not explicitly requested and exhibit unrealistic geometries (highlighted with red boxes). Similarly, in the image-to-image generation approach conditioned on segmentation masks, employing a more generic prompt results in the synthesis of implausible structures—such as slanted apartment-like buildings—demonstrating that prompt specificity directly impacts the semantic and geometric fidelity of the generated content.

5.3. Ablation on Filtering Threshold

As a final task, we investigate the sensitivity of our filtering scheme to the choice of strictness parameter, and we varied the k-value in the z-score threshold formulation, presented in Section 3.3. Specifically, we examined k = 1.5 , k = 2 , and k = 3 , corresponding to progressively stricter inclusion criteria for synthetic images based on their distance to real data clusters in feature space. These values were selected to reflect typical thresholds used in outlier detection based on standard deviations from the mean. A threshold of ± 3 σ standard deviations corresponds to roughly 99.7% coverage under the Gaussian curve, making it a common cutoff for identifying truly extreme values; in our case, outlier synthetic flood images are too dissimilar to real-word ones. Our results, shown in Table 7, demonstrate that a stricter threshold (i.e., k = 1.5 ) consistently yields better segmentation performance across the majority of evaluated models. This suggests that retaining only the most semantically aligned synthetic samples—those with the highest fidelity to real data distributions—is beneficial for generalization.
The performance gains are particularly evident in the convolutional models DeepLabV3 and FCN-ResNet50, and in both transformer-based SegFormer-B0 and Swin-T, which benefit significantly from the more selective filtering. This is likely due to the fact that these architectures possess deep and complex feature extractors, which are more sensitive to the domain discrepancies introduced by low-quality or semantically inconsistent synthetic data. By aggressively filtering out such samples, the learned representations remain more robust and transferable to real-world data. Interestingly, the intermediate threshold k = 2 shows the greatest benefit for the U-Net architecture. Unlike the others, U-Net has a more symmetric and shallow encoder–decoder structure, which may allow it to benefit from a slightly larger, more diverse synthetic training set—so long as the noise introduced remains within tolerable bounds. The less favorable performance of the rest of the examined models with k = 2 may highlight their greater sensitivity to noisy or out-of-distribution synthetic examples.
Table 7. Impact of the z-score threshold parameter k on the segmentation performance of different architectures trained on the filtered synthetic dataset. A lower k value corresponds to stricter filtering of synthetic images. For each method, the top-performing value per metric is shown in bold.
Table 7. Impact of the z-score threshold parameter k on the segmentation performance of different architectures trained on the filtered synthetic dataset. A lower k value corresponds to stricter filtering of synthetic images. For each method, the top-performing value per metric is shown in bold.
MethodThresh.
Par. (k)
Filt.
Imgs
Test Metrics
Acc (%)IoU (%)Pr (%)Rec (%)F1 (%)
DeepLabV3-0/58079.2160.9465.8989.0275.73
DeepLabV333/58079.1160.7665.8288.7775.59
DeepLabV3223/58078.6859.6965.7386.6774.76
DeepLabV31.553/58079.5561.3466.3489.0676.04
FCN-ResNet50-0/58083.7467.8270.8494.0980.83
FCN-ResNet5033/58083.5667.6470.5194.3280.70
FCN-ResNet50223/58082.8566.9069.2795.1480.17
FCN-ResNet501.553/58084.2268.1272.0892.5481.04
U-Net-0/58081.9165.5268.2194.3279.17
U-Net33/58084.6768.8272.6792.8581.53
U-Net223/58087.5772.8378.1791.4284.28
U-Net1.553/58086.9772.0076.8591.9483.72
SegFormer-B0-0/58080.1363.6165.6695.3377.76
SegFormer-B033/58080.6464.2366.2695.4578.22
SegFormer-B0223/58080.9264.3666.8394.5778.32
SegFormer-B01.553/58081.2264.9666.9995.4878.74
Swin-T-0/58083.5366.3972.1189.3279.80
Swin-T33/58082.1264.1370.4287.7878.15
Swin-T223/58081.9864.2869.8389.0078.26
Swin-T1.553/58084.8868.6973.6691.0681.44

5.4. Qualitative Segmentation Results

To evaluate the qualitative performance of the proposed framework, we present segmentation results from the best-performing model, namely U-Net, trained on the combined real-world and filtered synthetic dataset ( D r S D f i l t ). This configuration consistently achieved the highest F1-scores across validation and test sets, confirming its robustness and generalization capability under varying image conditions.
For interpretability and comprehensive assessment, representative segmentation outputs were selected based on the percentile sampling of the F1 score distribution over the test set. Specifically, we used the best-performing U-Net model variant, i.e., the variant trained with real and synthetic data ( D r S D f i l t ), and sorted all segmented test images in descending order of F1-scores and extracted five representative cases corresponding to the 0th (best), 20th, 40th, 60th, and 80th percentiles. This approach enables a structured visualization of the model’s performance across different levels of difficulty, from highly accurate segmentations to more challenging scenarios.
As illustrated in Figure 11, the top-ranked examples, e.g., 0th and 20th percentiles (Figure 11c,h), show high-performance segmentation with a near-precise delineation of flood boundaries and minimal false positives or negatives. These cases typically involve well-lit, high-contrast scenes with distinct flood regions and minimal occlusions. In contrast, lower percentile examples, e.g., 60th and 80th percentiles (Figure 11r,w), demonstrate the model’s limitations, often corresponding to visual ambiguities such as flooded vegetation, low contrast between land and water surfaces, or significant reflections and shadow artifacts. Nonetheless, even in these more difficult cases, the U-Net generally preserves the structural integrity of the flooded regions, indicating resilience to challenging input conditions.
These results further substantiate the effectiveness of combining real-world data with selectively filtered synthetic images, which likely enhanced the diversity and coverage of the training set, enabling the U-Net to generalize well across both typical and edge-case scenarios. For comparative reasons, we also present in Figure 11 the corresponding segmentation results of the other two models, FCN-ResNet50 and DeepLabV3, trained with the same configuration as the U-Net, together with the respective F1-scores achieved. FCN-ResNet50 performed second best and generally produced better results than the DeepLabV3 architecture, as observed in Figure 11d,i,s, respectively, compared to Figure 11e,j,t.
Similarly, Figure 12 displays the segmentation outcomes of the transformer-based models, Swin-T and SegFormer-B0, trained under identical configurations, alongside their respective F1-scores. In comparison to U-Net, both transformer models exhibit challenges in delineating object boundaries, often misclassifying non-flooded regions near flood boundaries as flooded. This is particularly evident in Figure 12i,j,n,o, when contrasted with the more accurate boundary segmentation of U-Net in (h) and (m). This behavior is partly attributed to the architectural characteristics of transformer models, such as the use of token-based representations and the shifted window mechanism in Swin-T, which can reduce spatial precision and impair the capture of fine-grained boundary details. Additionally, unlike U-Net, which incorporates strong inductive biases through convolutional operations and skip connections, transformer models lack inherent spatial locality and are more dependent on large-scale data to generalize effectively. However, there are cases such as Figure 12d,e,t, where transformer-based methods perform comparably to convolution-based ones, indicating the challenging nature of the task and the absence of a method that generalizes in every case. Overall, U-Net demonstrated superior detail preservation, which may be attributed to its stronger generalization capability under limited data conditions, whereas transformer-based architectures typically benefit from larger training sets and extended training time.
In Figure 13, we showcase the effectiveness of synthetic and semi-synthetic data for flood segmentation, presenting representative segmentation results from the same U-Net architecture trained under four distinct configurations: (i) using real-world images with manually annotated ground truth masks, (ii) using purely synthetic images generated via Stable Diffusion with automatically produced pseudo-labels, (iii) using semi-synthetic images created through image inpainting with corresponding pseudo-labels, and (iv) a combined dataset comprising real, synthetic, and semi-synthetic samples with their respective ground truth and pseudo-label masks. Five representative cases were chosen among the descending sorted differences of the F1-scores of the best- and-worst performing configurations, (iv) and (i), corresponding to the highest difference (first row), the differences of the 25th, 50th, 75th percentiles, and the lowest difference (last row).
Among the four cases, the combined dataset (case iv in the last column) achieved the highest performance in terms of segmentation accuracy, as measured by the F1-score. This indicates that the integration of real-world examples with synthetic and semi-synthetic data can significantly enhance the model’s generalization capability. The observed improvement is attributed to the increased variability and diversity introduced by the synthetic samples, which augment the training distribution and expose the model to a wider range of environmental conditions, textures, and structural layouts. This diversity appears to regularize the training and prevent overfitting to the limited real-world data.
Interestingly, both synthetic (case ii, third column) and semi-synthetic (case iii, fourth column) training independently produced segmentation results that were not only comparable to, but in some cases exceeded, the performance of models trained solely on real-world data (case i, second column), as observed in Figure 13c,d,n compared to Figure 13b,l. This highlights the capacity of U-Net to learn robust features even when trained exclusively on artificially generated data, provided that the pseudo-labels contain a sufficient region of interest despite their inherent noise. It also underscores the model’s resilience to imperfect supervision and demonstrates the potential of unsupervised or weakly supervised learning approaches for remote sensing tasks where high-quality annotated datasets are often scarce. The synthetic data, generated entirely from random noise using a generative diffusion process, provided a broad distribution of flood-like appearances, while the semi-synthetic data retained structural realism from the real imagery due to inpainting guided by true segmentation masks. Of course, there are also failure cases where complex patterns could not be learned and synthetic generated data as well as their combination with real data seemingly confused the model (see Figure 13w–y).
These findings collectively suggest that, in the absence of annotated datasets, the use of synthetic or semi-synthetic imagery in conjunction with automatic pseudo-labeling can offer a viable alternative for training deep segmentation models. Furthermore, combining such data with limited real-world samples results in a synergistic effect that further improves segmentation accuracy, advocating for hybrid training strategies in future work.

6. Conclusions

In this work, we presented a framework for constructing a synthetic dataset aimed at the task of flood segmentation in UAV imagery. Our approach integrates two distinct generative strategies: a text-to-image synthesis pipeline guided by flood-related textual prompts, and an image-to-image translation paradigm that leverages real segmentation masks as structural priors to generate realistic flood scenes via inpainting. These complementary approaches are unified to produce a diverse and semantically meaningful dataset.
To address the challenge of missing ground truth annotations for segmentation, we employ an unsupervised pseudo-labeling strategy to generate segmentation masks for the synthetic images. This allows us to construct paired image–mask samples without the need for manual annotation, significantly reducing the cost and effort typically associated with dataset curation. Furthermore, we incorporate a filtering stage based on outlier detection to ensure the realism and structural fidelity of the generated images, discarding samples that do not meet quality standards. To evaluate the effectiveness of the synthetic dataset, we conducted experiments with five top-performing flood segmentation models, three CNN and two transformer-based, assessing their performance on a real-world benchmark dataset. Our findings demonstrate that training with the synthetic data, even when annotated via unsupervised pseudo-labeling, not only leads to minor model performance drops, but the combination of synthetic and real-world data during training is able to improve model generalization and robustness. Our framework offers a scalable and low-cost solution for generating annotated flood segmentation datasets, with practical applications in disaster monitoring, remote sensing, and other vision-based environmental analysis tasks.
While our current framework provides a scalable solution for generating synthetic flood segmentation datasets, several promising directions remain open for future exploration. These include improving the accuracy of the unsupervised pseudo-labeling method by integrating stronger segmentation priors or combining multiple weak supervision signals. We also plan to explore domain adaptation techniques—such as adversarial training and feature alignment—to further reduce the gap between synthetic and real data. Lastly, we aim to extend the framework by considering multi-modal data synthesis for broader environmental monitoring applications.

Author Contributions

Conceptualization, G.S., K.B. and C.P.; Funding acquisition, C.P.; Investigation, G.S., K.B. and C.P.; Methodology, G.S., K.B. and C.P.; Software, G.S. and K.B.; Validation, G.S.; Supervision, K.B. and C.P.; Data Curation, G.S. and K.B.; Writing—original draft, G.S., K.B. and C.P.; Writing—review and editing, G.S., K.B. and C.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in this study are included in the article; further inquiries or requests can be directed to the corresponding authors.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ADNetAttentive Decoder Network
AIArtificial Intelligence
ASPPAtrous Spatial Pyramid Pooling
BiTBitemporal image Transformer
CNNConvolutional Neural Network
CVComputer Vision
DELTADeep Earth Learning, Tools, and Analysis
DLDeep Learning
DNNDeep Neural Network
ENetEfficient Neural Network
FADFlood Area Dataset
FCNFully Convolutional Network
FSSDFlood Semantic Segmentation Dataset
GANGenerative Adversarial Network
IoUIntersection over Union
MLMachine Learning
MMDMaximum Mean Discrepancy
MRFMarkov Random Field
NDWINormalized Difference Water Index
PCAPrincipal Component Analysis
PRPattern Recognition
ResNetResidual Network
SAMSegment Anything Model
SARSynthetic Aperture Radar
SegFormerSegmentation Transformer
SSSSSemi-Supervised Semantic Segmentation
UAVUnmanned Aerial Vehicle
UPerNetUnified Perceptual Parsing Network
VAEVariational Autoencoder
WSSSWeakly Supervised Semantic Segmentation

References

  1. Ritchie, H.; Rosado, P. Natural Disasters. 2022. Available online: https://ourworldindata.org/natural-disasters (accessed on 16 April 2025).
  2. Algiriyage, N.; Prasanna, R.; Stock, K.; Doyle, E.E.; Johnston, D. Multi-source multimodal data and deep learning for disaster response: A systematic review. SN Comput. Sci. 2022, 3, 92. [Google Scholar] [CrossRef] [PubMed]
  3. Linardos, V.; Drakaki, M.; Tzionas, P.; Karnavas, Y.L. Machine learning in disaster management: Recent developments in methods and applications. Mach. Learn. Knowl. Extr. 2022, 4, 446–473. [Google Scholar] [CrossRef]
  4. Albahri, A.; Khaleel, Y.L.; Habeeb, M.A.; Ismael, R.D.; Hameed, Q.A.; Deveci, M.; Homod, R.Z.; Albahri, O.; Alamoodi, A.; Alzubaidi, L. A systematic review of trustworthy artificial intelligence applications in natural disasters. Comput. Electr. Eng. 2024, 118, 109409. [Google Scholar] [CrossRef]
  5. Bacharidis, K.; Moirogiorgou, K.; Koukiou, G.; Giakos, G.; Zervakis, M. Stereo System for Remote Monitoring of River Flows. Multimed. Tools Appl. 2018, 77, 9535–9566. [Google Scholar] [CrossRef]
  6. Bentivoglio, R.; Isufi, E.; Jonkman, S.N.; Taormina, R. Deep learning methods for flood mapping: A review of existing applications and future research directions. Hydrol. Earth Syst. Sci. 2022, 26, 4345–4378. [Google Scholar] [CrossRef]
  7. Kyrkou, C.; Theocharides, T. Deep-Learning-Based Aerial Image Classification for Emergency Response Applications Using Unmanned Aerial Vehicles. In Proceedings of the CVPR Workshops, Long Beach, CA, USA, 16–20 June 2019; pp. 517–525. [Google Scholar]
  8. Zhu, X.; Liang, J.; Hauptmann, A. MSNet: A Multilevel Instance Segmentation Network for Natural Disaster Damage Assessment in Aerial Videos. arXiv 2020, arXiv:2006.16479. [Google Scholar]
  9. Zhu, X.; Liang, J.; Hauptmann, A. Msnet: A multilevel instance segmentation network for natural disaster damage assessment in aerial videos. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Online, 5–9 January 2021; pp. 2023–2032. [Google Scholar]
  10. Rahnemoonfar, M.; Chowdhury, T.; Sarkar, A.; Varshney, D.; Yari, M.; Murphy, R.R. Floodnet: A high resolution aerial imagery dataset for post flood scene understanding. IEEE Access 2021, 9, 89644–89654. [Google Scholar] [CrossRef]
  11. Karim, F.; Sharma, K.; Barman, N.R. Flood Area Segmentation. Available online: https://www.kaggle.com/datasets/faizalkarim/flood-area-segmentation (accessed on 6 January 2025).
  12. Hänsch, R.; Arndt, J.; Lunga, D.; Gibb, M.; Pedelose, T.; Boedihardjo, A.; Petrie, D.; Bacastow, T.M. Spacenet 8-the detection of flooded roads and buildings. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 1472–1480. [Google Scholar]
  13. Yang, L. Flood Semantic Segmentation Dataset. Available online: https://www.kaggle.com/datasets/lihuayang111265/flood-semantic-segmentation-dataset (accessed on 6 January 2025).
  14. Wieland, M.; Martinis, S.; Kiefl, R.; Gstaiger, V. Semantic segmentation of water bodies in very high-resolution satellite and aerial images. Remote Sens. Environ. 2023, 287, 113452. [Google Scholar] [CrossRef]
  15. Weber, E.; Papadopoulos, D.P.; Lapedriza, A.; Ofli, F.; Imran, M.; Torralba, A. Incidents1M: A large-scale dataset of images with natural disasters, damage, and incidents. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 4768–4781. [Google Scholar] [CrossRef]
  16. Rahnemoonfar, M.; Chowdhury, T.; Murphy, R. RescueNet: A high resolution UAV semantic segmentation dataset for natural disaster damage assessment. Sci. Data 2023, 10, 913. [Google Scholar] [CrossRef]
  17. Polushko, V.; Jenal, A.; Bongartz, J.; Weber, I.; Hatic, D.; Rösch, R.; März, T.; Rauhut, M.; Weinmann, A. BlessemFlood21: A High-Resolution Georeferenced Dataset for Advanced Analysis of River Flood Scenarios; IEEE Access: New York, NY, USA, 2024. [Google Scholar]
  18. Simantiris, G.; Panagiotakis, C. Unsupervised Color-Based Flood Segmentation in UAV Imagery. Remote Sens. 2024, 16, 2126. [Google Scholar] [CrossRef]
  19. Dong, Z.; Liang, Z.; Wang, G.; Amankwah, S.O.Y.; Feng, D.; Wei, X.; Duan, Z. Mapping inundation extents in Poyang Lake area using Sentinel-1 data and transformer-based change detection method. J. Hydrol. 2023, 620, 129455. [Google Scholar] [CrossRef]
  20. Drakonakis, G.I.; Tsagkatakis, G.; Fotiadou, K.; Tsakalides, P. OmbriaNet—Supervised flood mapping via convolutional neural networks using multitemporal sentinel-1 and sentinel-2 data fusion. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 2341–2356. [Google Scholar] [CrossRef]
  21. Hertel, V.; Chow, C.; Wani, O.; Wieland, M.; Martinis, S. Probabilistic SAR-based water segmentation with adapted Bayesian convolutional neural network. Remote Sens. Environ. 2023, 285, 113388. [Google Scholar] [CrossRef]
  22. Li, Z.; Demir, I. U-net-based semantic classification for flood extent extraction using SAR imagery and GEE platform: A case study for 2019 central US flooding. Sci. Total Environ. 2023, 869, 161757. [Google Scholar] [CrossRef] [PubMed]
  23. Hänsch, R.; Arndt, J.; Lunga, D.; Pedelose, T.; Boedihardjo, A.; Pfefferkorn, J.; Petrie, D.; Bacastow, T.M. SpaceNet 8: Winning Approaches to Multi-Class Feature Segmentation from Satellite Imagery for Flood Disasters. In Proceedings of the IGARSS 2023–2023 IEEE International Geoscience and Remote Sensing Symposium, Pasadena, CA, USA, 16–21 July 2023; pp. 1241–1244. [Google Scholar]
  24. Shastry, A.; Carter, E.; Coltin, B.; Sleeter, R.; McMichael, S.; Eggleston, J. Mapping floods from remote sensing data and quantifying the effects of surface obstruction by clouds and vegetation. Remote Sens. Environ. 2023, 291, 113556. [Google Scholar] [CrossRef]
  25. Chouhan, A.; Chutia, D.; Aggarwal, S.P. Attentive decoder network for flood analysis using sentinel 1 images. In Proceedings of the 2023 International Conference on Communication, Circuits, and Systems (IC3S), Bhubaneswar, India, 26–28 May 2023; pp. 1–5. [Google Scholar]
  26. nthizami, N.S.; Ma’sum, M.A.; Alhamidi, M.R.; Gamal, A.; Ardhianto, R.; Kurnianingsih; Jatmiko, W. Flood video segmentation on remotely sensed UAV using improved Efficient Neural Network. ICT Express 2022, 8, 347–351. [Google Scholar] [CrossRef]
  27. Yu, F.; Koltun, V. Multi-Scale Context Aggregation by Dilated Convolutions. arXiv 2016, arXiv:1511.07122. [Google Scholar]
  28. Şener, A.; Doğan, G.; Ergen, B. A novel convolutional neural network model with hybrid attentional atrous convolution module for detecting the areas affected by the flood. Earth Sci. Inform. 2024, 17, 193–209. [Google Scholar] [CrossRef]
  29. Wang, L.; Li, R.; Duan, C.; Zhang, C.; Meng, X.; Fang, S. A novel transformer based semantic segmentation scheme for fine-resolution remote sensing images. IEEE Geosci. Remote Sens. Lett. 2022, 19, 6506105. [Google Scholar] [CrossRef]
  30. Shi, L.; Yang, K.; Chen, Y.; Chen, G. An Interactive Prompt Based Network for Urban Floods Area Segmentation Using UAV Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 18, 935–948. [Google Scholar] [CrossRef]
  31. Landuyt, L.; Verhoest, N.E.; Van Coillie, F.M. Flood mapping in vegetated areas using an unsupervised clustering approach on sentinel-1 and-2 imagery. Remote Sens. 2020, 12, 3611. [Google Scholar] [CrossRef]
  32. Ibrahim, N.; Sharun, S.; Osman, M.; Mohamed, S.; Abdullah, S. The application of UAV images in flood detection using image segmentation techniques. Indones. J. Electr. Eng. Comput. Sci 2021, 23, 1219. [Google Scholar] [CrossRef]
  33. Bauer-Marschallinger, B.; Cao, S.; Tupas, M.E.; Roth, F.; Navacchi, C.; Melzer, T.; Freeman, V.; Wagner, W. Satellite-Based Flood Mapping through Bayesian Inference from a Sentinel-1 SAR Datacube. Remote Sens. 2022, 14, 3673. [Google Scholar] [CrossRef]
  34. McCormack, T.; Campanyà, J.; Naughton, O. A methodology for mapping annual flood extent using multi-temporal Sentinel-1 imagery. Remote Sens. Environ. 2022, 282, 113273. [Google Scholar] [CrossRef]
  35. Trombini, M.; Solarna, D.; Moser, G.; Dellepiane, S. A goal-driven unsupervised image segmentation method combining graph-based processing and Markov random fields. Pattern Recognit. 2023, 134, 109082. [Google Scholar] [CrossRef]
  36. Simantiris, G.; Panagiotakis, C. Unsupervised Deep Learning for Flood Segmentation in UAV imagery. In Proceedings of the 13th IAPR Workshop on Pattern Recognition in Remote Sensing, Kolkata, India, 1–4 December 2024. [Google Scholar]
  37. Ran, L.; Li, Y.; Liang, G.; Zhang, Y. Pseudo Labeling Methods for Semi-Supervised Semantic Segmentation: A Review and Future Perspectives. IEEE Trans. Circuits Syst. Video Technol. 2025, 35, 3054–3080. [Google Scholar] [CrossRef]
  38. Qin, Z.; Chen, Y.; Zhu, G.; Zhou, E.; Zhou, Y.; Zhou, Y.; Zhu, C. Enhanced Pseudo-Label Generation With Self-Supervised Training for Weakly- Supervised Semantic Segmentation. IEEE Trans. Circuits Syst. Video Technol. 2024, 34, 7017–7028. [Google Scholar] [CrossRef]
  39. Lu, X.; Jiang, Z.; Zhang, H. Weakly Supervised Remote Sensing Image Semantic Segmentation With Pseudo-Label Noise Suppression. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5406912. [Google Scholar] [CrossRef]
  40. Zou, Y.; Zhang, Z.; Zhang, H.; Li, C.L.; Bian, X.; Huang, J.B.; Pfister, T. Pseudoseg: Designing pseudo labels for semantic segmentation. arXiv 2020, arXiv:2010.09713. [Google Scholar]
  41. Paul, S.; Ganju, S. Flood segmentation on sentinel-1 SAR imagery with semi-supervised learning. arXiv 2021, arXiv:2107.08369. [Google Scholar]
  42. Li, J.; Meng, Y.; Li, Y.; Cui, Q.; Yang, X.; Tao, C.; Wang, Z.; Li, L.; Zhang, W. Accurate water extraction using remote sensing imagery based on normalized difference water index and unsupervised deep learning. J. Hydrol. 2022, 612, 128202. [Google Scholar] [CrossRef]
  43. He, Y.; Wang, J.; Zhang, Y.; Liao, C. An efficient urban flood mapping framework towards disaster response driven by weakly supervised semantic segmentation with decoupled training samples. ISPRS J. Photogramm. Remote Sens. 2024, 207, 338–358. [Google Scholar] [CrossRef]
  44. Kirillov, A.; Mintun, E.; Ravi, N.; Mao, H.; Rolland, C.; Gustafson, L.; Xiao, T.; Whitehead, S.; Berg, A.C.; Lo, W.Y.; et al. Segment Anything. arXiv 2023, arXiv:2304.02643. [Google Scholar]
  45. Savitha, G.; Girisha, S.; Sughosh, P.; Shetty, D.K.; Mymbilly Balakrishnan, J.; Paul, R.; Naik, N. Consistency Regularization for Semi-Supervised Semantic Segmentation of Flood Regions From SAR Images. IEEE Access 2025, 13, 9642–9653. [Google Scholar] [CrossRef]
  46. Croitoru, F.A.; Hondru, V.; Ionescu, R.T.; Shah, M. Diffusion models in vision: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 10850–10869. [Google Scholar] [CrossRef]
  47. Rombach, R.; Blattmann, A.; Lorenz, D.; Esser, P.; Ommer, B. High-resolution image synthesis with latent diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 10684–10695. [Google Scholar]
  48. Dhariwal, P.; Nichol, A. Diffusion models beat gans on image synthesis. Adv. Neural Inf. Process. Syst. 2021, 34, 8780–8794. [Google Scholar]
  49. Song, J.; Meng, C.; Ermon, S. Denoising diffusion implicit models. arXiv 2020, arXiv:2010.02502. [Google Scholar]
  50. Saharia, C.; Chan, W.; Saxena, S.; Li, L.; Whang, J.; Denton, E.L.; Ghasemipour, K.; Gontijo Lopes, R.; Karagol Ayan, B.; Salimans, T.; et al. Photorealistic text-to-image diffusion models with deep language understanding. Adv. Neural Inf. Process. Syst. 2022, 35, 36479–36494. [Google Scholar]
  51. Batzolis, G.; Stanczuk, J.; Schönlieb, C.B.; Etmann, C. Conditional image generation with score-based diffusion models. arXiv 2021, arXiv:2111.13606. [Google Scholar]
  52. Saharia, C.; Chan, W.; Chang, H.; Lee, C.; Ho, J.; Salimans, T.; Fleet, D.; Norouzi, M. Palette: Image-to-image diffusion models. In Proceedings of the ACM SIGGRAPH 2022 Conference Proceedings, Vancouver, BC, Canada, 7–11 August 2022; pp. 1–10. [Google Scholar]
  53. Wang, T.; Zhang, T.; Zhang, B.; Ouyang, H.; Chen, D.; Chen, Q.; Wen, F. Pretraining is all you need for image-to-image translation. arXiv 2022, arXiv:2205.12952. [Google Scholar]
  54. Ramesh, A.; Pavlov, M.; Goh, G.; Gray, S.; Voss, C.; Radford, A.; Chen, M.; Sutskever, I. Zero-shot text-to-image generation. In Proceedings of the International Conference on Machine Learning, Online, 18–24 July 2021; pp. 8821–8831. [Google Scholar]
  55. Ramesh, A.; Dhariwal, P.; Nichol, A.; Chu, C.; Chen, M. Hierarchical text-conditional image generation with clip latents. arXiv 2022, arXiv:2204.06125. [Google Scholar]
  56. Müller-Franzes, G.; Niehues, J.M.; Khader, F.; Arasteh, S.T.; Haarburger, C.; Kuhl, C.; Wang, T.; Han, T.; Nolte, T.; Nebelung, S.; et al. A multimodal comparison of latent denoising diffusion probabilistic models and generative adversarial networks for medical image synthesis. Sci. Rep. 2023, 13, 12098. [Google Scholar] [CrossRef] [PubMed]
  57. Vivekananthan, S. Comparative analysis of generative models: Enhancing image synthesis with vaes, gans, and stable diffusion. arXiv 2024, arXiv:2408.08751. [Google Scholar]
  58. Sarfraz, S.; Sharma, V.; Stiefelhagen, R. Efficient parameter-free clustering using first neighbor relations. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 8934–8943. [Google Scholar]
  59. Zotou, P.; Bacharidis, K.; Argyros, A. Leveraging FINCH and K-means for Enhanced Cluster- Based Instance Selection. In Proceedings of the 18th European Conference on Computer Vision (ECCV), Milan, Italy, 29 September–4 October 2024. [Google Scholar] [CrossRef]
  60. Chen, L.C.; Papandreou, G.; Schroff, F.; Adam, H. Rethinking atrous convolution for semantic image segmentation. arXiv 2017, arXiv:1706.05587. [Google Scholar]
  61. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  62. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; Part III 18. Springer: Berlin/Heidelberg, Germany; pp. 234–241. [Google Scholar]
  63. Xie, E.; Wang, W.; Yu, Z.; Anandkumar, A.; Alvarez, J.M.; Luo, P. SegFormer: Simple and efficient design for semantic segmentation with transformers. Adv. Neural Inf. Process. Syst. 2021, 34, 12077–12090. [Google Scholar]
  64. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  65. Schuhmann, C.; Beaumont, R.; Vencu, R.; Gordon, C.; Wightman, R.; Cherti, M.; Coombes, T.; Katta, A.; Mullis, C.; Wortsman, M.; et al. Laion-5b: An open large-scale dataset for training next generation image-text models. Adv. Neural Inf. Process. Syst. 2022, 35, 25278–25294. [Google Scholar]
Figure 1. Flood-related aerial image datasets and their sizes (in images). Each bubble has its own color, except for gray bubbles that indicate the absence of explicit flood annotations. Sizes for AIDER and Incidents1M reflect only flood-related images.
Figure 1. Flood-related aerial image datasets and their sizes (in images). Each bubble has its own color, except for gray bubbles that indicate the absence of explicit flood annotations. Sizes for AIDER and Incidents1M reflect only flood-related images.
Sensors 25 03586 g001
Figure 2. Schematic overview of the proposed framework. Synthetic and semi-synthetic images are generated via Stable Diffusion using prompt engineering, with the latter incorporating original ground truth labels (GT) through image inpainting. Pseudo-labels for original (I), synthetic, and semi-synthetic images are produced via unsupervised segmentation. Outlier filtering is applied, and the combined datasets are used to train deep neural networks for flood segmentation.
Figure 2. Schematic overview of the proposed framework. Synthetic and semi-synthetic images are generated via Stable Diffusion using prompt engineering, with the latter incorporating original ground truth labels (GT) through image inpainting. Pseudo-labels for original (I), synthetic, and semi-synthetic images are produced via unsupervised segmentation. Outlier filtering is applied, and the combined datasets are used to train deep neural networks for flood segmentation.
Sensors 25 03586 g002
Figure 3. Overview of the synthetic dataset creation pipeline. Synthetic flood images are generated via (i) text-to-image synthesis and (ii) image inpainting guided by real segmentation masks (GT) from real images (Ir). Unsupervised pseudo-labeling (PL Gen) yields segmentation masks PLs and PLip, for synthetic (Is) and semi-synthetic (Iip) images, respectively. The resulting dataset (SDall) is refined through feature-based clustering to remove outliers, producing the final synthetic dataset (SDfilt) used for training.
Figure 3. Overview of the synthetic dataset creation pipeline. Synthetic flood images are generated via (i) text-to-image synthesis and (ii) image inpainting guided by real segmentation masks (GT) from real images (Ir). Unsupervised pseudo-labeling (PL Gen) yields segmentation masks PLs and PLip, for synthetic (Is) and semi-synthetic (Iip) images, respectively. The resulting dataset (SDall) is refined through feature-based clustering to remove outliers, producing the final synthetic dataset (SDfilt) used for training.
Sensors 25 03586 g003
Figure 4. Illustration of the image generation process using Stable Diffusion over 50 denoising steps. The sequence shows the initial random noise and intermediate outputs at every 10th step (t) until the final synthesized image. Diversity in the generated outputs is achieved through carefully constructed text prompts. The model successfully generates realistic flooded environments, including urban/peri-urban scenes (a,b), and rural landscapes (c,d). Variations in sky presence which can include sky (b,d) or not (a,c), simulate different camera viewpoints and orientations.
Figure 4. Illustration of the image generation process using Stable Diffusion over 50 denoising steps. The sequence shows the initial random noise and intermediate outputs at every 10th step (t) until the final synthesized image. Diversity in the generated outputs is achieved through carefully constructed text prompts. The model successfully generates realistic flooded environments, including urban/peri-urban scenes (a,b), and rural landscapes (c,d). Variations in sky presence which can include sky (b,d) or not (a,c), simulate different camera viewpoints and orientations.
Sensors 25 03586 g004aSensors 25 03586 g004b
Figure 5. Synthetic images I s (a) from Figure 4 representing flooded urban/peri-urban environments with sky absence and presence, and rural scenes also without and with sky. Real-world UAV captured flood-related images with the same scene diversity (c), along with their respective ground truths (d), which are used, as described in Section 3.2, to generate semi-synthetic images I i p (e) via inpainting. Respective pseudo-labels, PLs and PLip, are overlaid in blue for the synthetic Is (b) and semi-synthetic images Iip (f).
Figure 5. Synthetic images I s (a) from Figure 4 representing flooded urban/peri-urban environments with sky absence and presence, and rural scenes also without and with sky. Real-world UAV captured flood-related images with the same scene diversity (c), along with their respective ground truths (d), which are used, as described in Section 3.2, to generate semi-synthetic images I i p (e) via inpainting. Respective pseudo-labels, PLs and PLip, are overlaid in blue for the synthetic Is (b) and semi-synthetic images Iip (f).
Sensors 25 03586 g005
Figure 6. Representative examples of synthetic (a) and semi-synthetic images (d) which were filtered with outlier threshold k = 1.5 . Semi-synthetic images I i p were generated as described in Section 3.2, with the ground truth G T (c) used for in-painting derived from the real-world UAV captured flood-related images I r (b).
Figure 6. Representative examples of synthetic (a) and semi-synthetic images (d) which were filtered with outlier threshold k = 1.5 . Semi-synthetic images I i p were generated as described in Section 3.2, with the ground truth G T (c) used for in-painting derived from the real-world UAV captured flood-related images I r (b).
Sensors 25 03586 g006
Figure 7. Sample images from the training dataset FAD [11] (a) and the corresponding ground truths (b), which were used in Section 3.2 to generate semi-synthetic images, and from the test dataset FSSD [13] (c) together with their respective ground truths (d).
Figure 7. Sample images from the training dataset FAD [11] (a) and the corresponding ground truths (b), which were used in Section 3.2 to generate semi-synthetic images, and from the test dataset FSSD [13] (c) together with their respective ground truths (d).
Sensors 25 03586 g007
Figure 8. The F1-score of DeepLabV3, FCN-ResNet50, U-Net, SegFormer-B0, and Swin-T on the test dataset (FSSD) according to training with the real dataset with ground truth ( D r ) and pseudo-label ( D p l ) masks, text-to-image synthetic dataset ( S D s ), the semi-synthetic image inpainting dataset ( S D i p ), and their union, unfiltered ( D r S D a l l ) and filtered ( D r S D f i l t ).
Figure 8. The F1-score of DeepLabV3, FCN-ResNet50, U-Net, SegFormer-B0, and Swin-T on the test dataset (FSSD) according to training with the real dataset with ground truth ( D r ) and pseudo-label ( D p l ) masks, text-to-image synthetic dataset ( S D s ), the semi-synthetic image inpainting dataset ( S D i p ), and their union, unfiltered ( D r S D a l l ) and filtered ( D r S D f i l t ).
Sensors 25 03586 g008
Figure 9. Two-dimensional PCA visualization with blue, red and green colors of ResNet50 feature embeddings extracted from real (Dr) and synthetic (SDs, SDip) datasets, respectively, with the best fit ellipses of each dataset calculated by 2D normal distribution fitting. Each point represents an image, colored by dataset. The substantial overlap suggests strong visual alignment between real and synthetic domains.
Figure 9. Two-dimensional PCA visualization with blue, red and green colors of ResNet50 feature embeddings extracted from real (Dr) and synthetic (SDs, SDip) datasets, respectively, with the best fit ellipses of each dataset calculated by 2D normal distribution fitting. Each point represents an image, colored by dataset. The substantial overlap suggests strong visual alignment between real and synthetic domains.
Sensors 25 03586 g009
Figure 10. Comparison of accepted (top) and rejected (bottom) synthetic flood images generated via prompt-to-image (left) and image-to image (right) translation. Problematic regions—including artifacts, erroneous configurations, and undesired objects—are highlighted with red bounding boxes.
Figure 10. Comparison of accepted (top) and rejected (bottom) synthetic flood images generated via prompt-to-image (left) and image-to image (right) translation. Problematic regions—including artifacts, erroneous configurations, and undesired objects—are highlighted with red bounding boxes.
Sensors 25 03586 g010
Figure 11. Original images (first column (a,f,k,p,u)), ground truth masks (second column (b,g,l,q,v)), and representative segmentation results of the U-Net (third column), FCN-ResNet50 (fourth column), and DeepLabV3 (fifth column) from the FSSD test dataset. The masks are overlaid in blue. Rows correspond to the 0 % (best results), 20 % , 40 % , 60 % , and 80 % percentile of the descending sorted F1-score values of the top-performing CNN architecture, U-Net, trained with real and synthetic data ( D r S D f i l t ). FCN-ResNet50 and DeepLabV3 were also trained under the same conditions.
Figure 11. Original images (first column (a,f,k,p,u)), ground truth masks (second column (b,g,l,q,v)), and representative segmentation results of the U-Net (third column), FCN-ResNet50 (fourth column), and DeepLabV3 (fifth column) from the FSSD test dataset. The masks are overlaid in blue. Rows correspond to the 0 % (best results), 20 % , 40 % , 60 % , and 80 % percentile of the descending sorted F1-score values of the top-performing CNN architecture, U-Net, trained with real and synthetic data ( D r S D f i l t ). FCN-ResNet50 and DeepLabV3 were also trained under the same conditions.
Sensors 25 03586 g011
Figure 12. Original images (first column (a,f,k,p,u)), ground truth masks (second column (b,g,l,q,v)), and representative segmentation results of the Swin-T (third column), and SegFormer-B0 (fourth column) from the FSSD test dataset. The masks are overlaid in blue. Rows correspond to the 0 % (best results), 20 % , 40 % , 60 % and 80 % percentile of the descending sorted F1-score values of the top-performing overall model, U-Net, trained with real and synthetic data ( D r S D f i l t ). Swin-T and SegFormer-B0 were also trained under the same conditions.
Figure 12. Original images (first column (a,f,k,p,u)), ground truth masks (second column (b,g,l,q,v)), and representative segmentation results of the Swin-T (third column), and SegFormer-B0 (fourth column) from the FSSD test dataset. The masks are overlaid in blue. Rows correspond to the 0 % (best results), 20 % , 40 % , 60 % and 80 % percentile of the descending sorted F1-score values of the top-performing overall model, U-Net, trained with real and synthetic data ( D r S D f i l t ). Swin-T and SegFormer-B0 were also trained under the same conditions.
Sensors 25 03586 g012
Figure 13. Original images (first column (a,f,k,p,u)) and representative segmentation results from the FSSD test dataset of the U-Net architecture trained with different datasets: real-world images and ground truths (second column), synthetic images and corresponding pseudo-label masks (third column), semi-synthetic images and corresponding pseudo-label masks (fourth column), and combined real-world data and filtered synthetic datasets and corresponding ground truth and pseudo-labels (fifth column). Segmentations are overlaid in blue. Rows correspond to the 0 % (best results), 25 % , 50 % , 75 % and 100 % (worst results) percentile of the descending sorted F1-score differences of the top performing U-Net architecture (last columns) and its counterpart trained with real-world images and ground truth (second column).
Figure 13. Original images (first column (a,f,k,p,u)) and representative segmentation results from the FSSD test dataset of the U-Net architecture trained with different datasets: real-world images and ground truths (second column), synthetic images and corresponding pseudo-label masks (third column), semi-synthetic images and corresponding pseudo-label masks (fourth column), and combined real-world data and filtered synthetic datasets and corresponding ground truth and pseudo-labels (fifth column). Segmentations are overlaid in blue. Rows correspond to the 0 % (best results), 25 % , 50 % , 75 % and 100 % (worst results) percentile of the descending sorted F1-score differences of the top performing U-Net architecture (last columns) and its counterpart trained with real-world images and ground truth (second column).
Sensors 25 03586 g013
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Simantiris, G.; Bacharidis, K.; Panagiotakis, C. Closing the Domain Gap: Can Pseudo-Labels from Synthetic UAV Data Enable Real-World Flood Segmentation? Sensors 2025, 25, 3586. https://doi.org/10.3390/s25123586

AMA Style

Simantiris G, Bacharidis K, Panagiotakis C. Closing the Domain Gap: Can Pseudo-Labels from Synthetic UAV Data Enable Real-World Flood Segmentation? Sensors. 2025; 25(12):3586. https://doi.org/10.3390/s25123586

Chicago/Turabian Style

Simantiris, Georgios, Konstantinos Bacharidis, and Costas Panagiotakis. 2025. "Closing the Domain Gap: Can Pseudo-Labels from Synthetic UAV Data Enable Real-World Flood Segmentation?" Sensors 25, no. 12: 3586. https://doi.org/10.3390/s25123586

APA Style

Simantiris, G., Bacharidis, K., & Panagiotakis, C. (2025). Closing the Domain Gap: Can Pseudo-Labels from Synthetic UAV Data Enable Real-World Flood Segmentation? Sensors, 25(12), 3586. https://doi.org/10.3390/s25123586

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop