Next Article in Journal
Telecom Fraud Detection Based on Large Language Models: A Multi-Role, Multi-Layer Prompting Strategy
Previous Article in Journal
Improved PPIM—A Method to Further Improve the Measurement Accuracy of the Cross-Sectional Area of the Conveying Material Load
Previous Article in Special Issue
OpenBeePose: A Remote Sensing Framework for Bee Pose Estimation Using Deep Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hybrid Object-Based Augmentation and Histogram Matching for Cross-Domain Building Segmentation in Remote Sensing

Department of AI Computer Engineering, Far East University, Eumseong 27601, Republic of Korea
*
Author to whom correspondence should be addressed.
Appl. Sci. 2026, 16(1), 543; https://doi.org/10.3390/app16010543
Submission received: 8 December 2025 / Revised: 26 December 2025 / Accepted: 2 January 2026 / Published: 5 January 2026

Abstract

Cross-domain building segmentation in high-resolution remote sensing imagery underpins urban change monitoring, disaster assessment, and exposure mapping. However, differences in sensors, regions, and imaging conditions create structural and radiometric domain gaps that degrade model generalization. Most existing methods adopt model-centric domain adaptation with additional networks or losses, complicating training and deployment. We propose a data-centric framework, Hybrid Object-Based Augmentation and Histogram Matching (Hybrid OBA–HM), which improves cross-domain building segmentation without modifying the backbone architecture or using target-domain labels. The proposed framework comprises two stages: (i) object-based augmentation to increase structural diversity and building coverage, and (ii) histogram-based normalization to mitigate radiometric discrepancies across domains. Experiments on OpenEarthMap and cross-city transfer among three KOMPSAT-3A scenes show that Hybrid OBA–HM improves F1-scores from 0.808 to 0.840 and from 0.455 to 0.652, respectively, while maintaining an object-level intersection over union of 0.89 for replaced buildings. Domain-indicator analysis further reveals larger gains under stronger radiometric and geometric mismatches, indicating that the proposed framework strengthens cross-domain generalization and provides practical guidance by relating simple domain diagnostics (e.g., brightness/color and orientation mismatch indicators) to the expected benefits of augmentation and normalization when adapting to new domains.

1. Introduction

Buildings extracted from high-resolution remote sensing imagery provide fundamental information for characterizing the human-built environment from local to global scales. Building maps are essential for various geospatial applications, including urban change monitoring, disaster management, map updating, and smart city development. Recent deep learning-based building segmentation methods have demonstrated strong performance on various benchmark datasets. However, models trained on specific regions or sensors often suffer from substantial performance degradation when applied to images from different regions, sensors, or acquisition conditions. Cross-domain building segmentation, which seeks to maintain consistent performance under such domain shifts, therefore remains a core challenge in the field of remote sensing, and the demand for robust automatic extraction of reliable building information from high-resolution imagery continues to grow.
Recent work in change detection and map updating has introduced fully transformer-based U-Net frameworks [1] that integrate aerial imagery, digital surface models (DSMs), and vector information; self-supervised models that jointly address change detection and damage assessment; and compressed-sensing-based strategies that reduce feature dimensionality for efficient map updating [2,3,4]. Complementary studies have also proposed fusion models combining GIS maps and high-resolution satellite images to generate accurate building outline information, as well as automatic building-outline extraction pipelines for large-scale map processing [5,6]. Building semantic segmentation further plays a critical role in disaster response by enabling rapid mapping of the spatial distribution of damaged buildings and supporting timely damage assessment across diverse scenarios [7,8,9,10,11,12,13]. In urban planning and smart city applications, its value for urban analysis, urban management, and planning decision-making has been demonstrated through GIS-based high-precision building-outline generation frameworks, building footprints derived from LiDAR and very-high-resolution (VHR) imagery, and segmentation models designed for complex backgrounds [5,14,15].
However, building segmentation models typically perform best when the target dataset is statistically similar to the training data. When the target dataset differs from the training dataset in sensor type, acquisition time, season, weather conditions, or other factors, performance often deteriorates due to domain gaps between source and target distributions; this issue is amplified in satellite and aerial imagery, where building shape, scale, and materials vary substantially across regions. Domain adaptation aims to improve target-domain performance by reducing these distributional differences between a labeled source domain and an unlabeled target domain.
In remote sensing semantic segmentation, Yu et al. categorized domain adaptation methods into discrepancy-based, adversarial, and pseudo-label/self-training approaches [16]. Discrepancy-based methods compensate for cross-domain differences in class distribution and class sets via distribution matching and clustering strategies, or by identifying common, additional, and missing classes [17,18]. Adversarial approaches use discriminators and generative adversarial networks (GANs) to align global and class-specific distributions in feature and output spaces, and have been applied to maintain structural consistency and reduce inter-city domain gaps through output-space alignment, multi-stage adversarial learning, and scale-aware discrimination [19,20,21]. Related GAN-based domain adaptation studies have improved segmentation and cloud detection performance by mitigating spectral and resolution differences between cities and sensors through frequency–space transformation, color–spectrum mapping, semantic-label-based reconstruction, and cycle-consistent style transformation [22,23,24,25,26]. Finally, pseudo-labeling and self-training strategies exploit target-domain predictions to improve generalization under high class asymmetry and inter-city domain heterogeneity, including multi-source knowledge integration, dynamic co-learning frameworks, and image-level weakly supervised domain adaptation [27,28,29].
These domain adaptation studies have made significant progress in reducing domain gaps. They do so by minimizing distributional mismatch, aligning styles and features through adversarial learning, and leveraging self-training with pseudo-labels and weak supervision. Nevertheless, many approaches still rely primarily on image-level statistics or style alignment. As a result, structural (geometric) differences—such as building shape, size, and density—are often handled only implicitly, even though they can vary substantially with geographic region, sensor type, and acquisition conditions [16,17,18]. Pixel-level style transfer and related adversarial frameworks (e.g., CycleGAN- or ColorMapGAN-based methods) can effectively mitigate texture differences. However, several studies report that these methods may distort building boundaries or roof structures during training [23,25]. Pseudo-labeling and self-learning techniques help alleviate the lack of target-domain labels. Yet, when the domain gap is large, pseudo-label errors can accumulate during iterative training, and class asymmetry or background semantic shifts can further degrade performance [27,30,31,32]. Moreover, remote sensing imagery exhibits wide geographic, seasonal, and sensor diversity. This diversity makes it difficult to sufficiently represent shifts in broader scene context, such as land-cover patterns and backgrounds, using a single alignment mechanism [28,32,33]. Finally, many studies emphasize aggregate metrics such as mean intersection-over-union (mIoU) and F1-score. In contrast, relatively few provide interpretive analyses that explain how specific domain mismatches (e.g., color, structure, and context) relate to performance changes [16,34].
Given the limitations of existing domain adaptation studies, building segmentation in remote sensing requires a data-centric strategy that can simultaneously address (i) geometric diversity (e.g., building shape, area, layout, and orientation) and (ii) radiometric characteristics (illumination and color). This need arises because, when the region, sensor, or imaging conditions change, variations in building structure and background scene context occur together, making it difficult for simple style-transfer or feature-alignment schemes alone to capture such multidimensional domain shifts.
To better cope with structural discrepancies, object-based augmentation (OBA) has been explored as a way to directly expand structural diversity by manipulating training samples. Prior work has reported that synthesizing extracted objects onto diverse backgrounds, cut-and-paste operations, and adversarial instance insertion can improve performance in semantic segmentation and building change detection tasks [35,36,37,38,39]. In addition, the effects of combining geometric and photometric augmentations have been examined for SAR-based building footprint segmentation [40]. However, most OBA-oriented studies primarily expand object distributions within a single sensor and region, which limits their ability to reduce domain gaps or explicitly interpret coupled changes in structure and scene context under cross-domain settings with different sensors, regions, and imaging conditions.
In parallel, histogram- or statistics-based normalization provides a simple yet effective data-driven approach for mitigating domain shift by directly aligning pixel-intensity distributions. Examples such as randomized histogram matching, multi-center/multi-vendor histogram matching, photometric alignment, and class-wise histogram matching suggest that global- or class-level statistical alignment can reduce performance degradation across sensors and scenes [41,42,43,44]. Nonetheless, these techniques largely focus on source-to-target radiometric alignment and do not explicitly integrate object-level synthetic sampling to jointly bridge structural and radiometric gaps.
Based on this motivation, this study proposes Hybrid OBA–HM, a hybrid framework that combines OBA (to expand object-level structural diversity) with histogram matching (HM) (to perform global radiometric normalization). The proposed approach aims to simultaneously reduce inter-domain differences in luminance/color and geometry. To minimize structural distortion and mitigate side effects of global transformations often observed in GAN-based style transfer or feature-alignment-based adaptation, the method analyzes statistical characteristics of building objects and backgrounds separately, while applying the final radiometric transformation at the entire-image level.
Existing object-based augmentation methods expand the distribution of building instances, but they are often tuned to a single sensor or region and may provide limited control over local context consistency and tile-level building coverage in cross-domain settings [35,36,37]. In contrast, histogram- or statistics-based normalization aligns global intensity distributions but cannot capture geometric shifts such as changes in building density, size, or dominant orientation. Hybrid OBA–HM therefore combines complementary mechanisms: object-based augmentation increases plausible structural diversity and coverage, while histogram matching regularizes brightness and color toward the target domain, jointly mitigating the coupled structure–appearance gap in cross-domain building segmentation.
Furthermore, to systematically verify the effectiveness of the proposed technique under various source–target combinations using OEM global and KOMPSAT-3A regional scenarios, we introduce domain indicators that help characterize domain gaps and analyze their association with segmentation performance. The main contributions of this study are summarized as follows:
  • We propose a hybrid data-centric framework that integrates object-based augmentation with histogram matching to jointly mitigate geometric and radiometric domain gaps in cross-domain building segmentation.
  • We design an object-based augmentation mechanism that increases structural diversity and building coverage through multi-stage object replacement, object-count-based coverage filtering, and geometric transformations, while preserving object-level geometric consistency across domains.
  • We introduce domain indicators based on brightness, color, and directionality and propose an analysis protocol that quantitatively relates domain discrepancies to segmentation behavior across transfer scenarios.
The remainder of this paper is organized as follows. Section 2 describes the study areas and datasets and details the proposed Hybrid OBA–HM framework, including its object-based augmentation and histogram-based normalization components. Section 3 presents the experimental setup, parameter configuration, and quantitative and qualitative evaluation results for global and regional cross-domain scenarios. Section 4 discusses the results and their implications, and Section 5 concludes the paper and outlines directions for future research. Additional details supporting clarity and reproducibility (notation, key settings, dataset splits, and extended results) are provided in Appendix A (Table A1, Table A2, Table A3, Table A4 and Table A5).

2. Materials and Methods

Figure 1 presents a flowchart of the proposed Hybrid OBA–HM pipeline. The workflow is organized into three main modules. (a) The Object-Based Augmentation (OBA) module increases building-level structural diversity via object replacement, generating an OBA-augmented dataset. (b) The Histogram-Based Normalization and Enhanced Geometric Augmentation (HM + EGA) module takes the OBA-augmented data as input, performs radiometric alignment using EMD-based histogram analysis and Shannon entropy-based target selection, applies global histogram matching, and then conducts enhanced geometric augmentation to obtain an HM–EGA-augmented dataset. (c) The final training set is formed by combining the original images with the augmented data and is used to train a U-Net–VGG19 segmentation model; the trained model is subsequently evaluated for cross-domain generalization on OpenEarthMap and KOMPSAT-3A scenarios.

2.1. Study Areas and Datasets

The proposed Hybrid (OBA + HM) framework was evaluated on two remote sensing building segmentation datasets: the OpenEarthMap (OEM) dataset [45] for global generalization experiments and a KOMPSAT-3A dataset for regional generalization. The OEM dataset consists of high-resolution satellite and aerial imagery with spatial resolutions ranging from 0.25 m to 0.5 m, collected from 97 regions across six continents. This dataset, created by merging existing datasets from various continental regions, exhibits differences in acquisition time, weather conditions, and sensor characteristics. Consequently, regional variations in brightness and color distributions occur, as well as geometric heterogeneity resulting from differences in building architecture and urban density. The KOMPSAT-3A dataset comprises images of three South Korean cities—Seoul, Incheon, and Jeju—captured by the same optical satellite, KOMPSAT-3A. The satellite provides a panchromatic resolution of 0.55 m and a multispectral resolution of 2.2 m. In this study, we used vendor-provided pan-sharpened (PS) imagery, which yields an effective spatial resolution of approximately 0.5 m. The specific pansharpening (fusion) algorithm used to generate the PS product is not specified in the product documentation. For training, we used the PS RGB bands and converted the original 16-bit imagery to 8-bit prior to dataset construction. The dataset exhibits significant inter-city variations in brightness and color distributions, as well as geometric heterogeneity caused by differences in building shape, roof orientation, and density.
Figure 2 shows representative examples of input images from the OEM and KOMPSAT-3A datasets. A European scene from the OEM dataset displays mainly isolated buildings, whereas the KOMPSAT-3A scenes over Seoul, Incheon, and Jeju depict closely packed buildings with strong directional layout patterns, highlighting both radiometric and geometric discrepancies between the two domains.

2.2. Object-Based Data Augmentation

The objective of this study is to improve the generalization performance of a building segmentation model by reducing domain discrepancies caused by differences in region, sensor characteristics, and illumination conditions. The proposed method combines OBA and histogram-based normalization (HM) within a unified two-stage framework. In the first stage, the object-based data augmentation performs building-level object replacement, geometric alignment, and transformation to obtain greater structural diversity within the training dataset. In the second stage, the histogram-based normalization reduces photometric distribution differences between the training and target datasets. After histogram alignment, geometric transformations are further applied to the normalized data to enhance model robustness. Through this two-stage process, the proposed method strengthens the robustness and adaptability of the building segmentation model against combined photometric and geometric variations across domains, thereby improving cross-domain generalization performance.
For clarity, the proposed augmentation workflow is organized into two sequential stages. Stage 1 (Object-Based Augmentation, OBA) increases structural diversity by replacing building instances within their original footprints and applying geometric transformations. Stage 2 (Histogram Matching with Enhanced Geometric Augmentation, HM + EGA) reduces radiometric mismatch to the target domain via histogram matching and then reapplies geometric transformations to improve robustness. The details of Stage 1 are described in Section 2.2, and Stage 2 is presented in Section 2.3.
Existing object-based augmentation research has improved semantic segmentation and change detection performance by synthesizing building or change objects from label masks onto various backgrounds or by applying cut-and-paste and adversarial instance insertion [35,36,37,38,39,40]. Specifically, ref. [35] is conceptually related to our work because it leverages object-level manipulation to enrich training data. However, our Stage-1 design differs from [35] in three key respects. First, ref. [35] selects a background image and places independent object instances onto it. In contrast, we keep the original background fixed and replace a building instance with another geometrically similar building instance within the same scene context. This design preserves local layout consistency. Second, ref. [35] assumes separable backgrounds with sufficient free space, which can be limiting in dense urban scenes where buildings occupy most of the image. Our replacement-based design does not require large empty background regions, making it suitable for building-dense imagery. Third, our method operates at a substantially larger scale. In our experiments, the number of replaced building instances ranges from approximately 38,000 to 87,000 per scenario. By comparison, ref. [35] reports far fewer processed instances (below 1000 in the training set). However, most existing approaches operate within a single sensor or single region. They also rely on random object replacement and simple geometric or color transformations. This limits their ability to control object-level geometric consistency and tile-level building coverage in cross-domain settings. The Stage-1 OBA module in this study addresses these limitations by combining building-polygon-based candidate selection, coverage filtering, and primary geometric augmentation. The module generates cross-domain–oriented augmented samples in which building structure and density are not unrealistically distorted, even when the source and target domains differ.
The proposed pipeline uses the original images, label images, and the polygonal building regions extracted from the label masks as inputs. The sequence of operations is as follows:
(1)
polygon extraction and attribute computation;
(2)
multi-stage object replacement;
(3)
object-count-based coverage filtering;
(4)
first-stage geometric augmentation;
(5)
histogram-based normalization; and
(6)
enhanced geometric augmentation.
The output dataset is constructed by combining the augmented images and labels generated after the enhanced geometric augmentation stage with the original images and label maps, thereby producing a comprehensive training dataset for model learning.
The proposed method enhances object-level structural diversity from a limited training dataset while normalizing the photometric distribution of training images. Through this combination, the method effectively mitigates the degradation of model performance caused by domain discrepancies and achieves greater generalization capability across diverse regions and sensors.

2.2.1. Polygon Extraction and Attribute Computation

The first step in object-based data augmentation is to extract building polygons from the label images of the training dataset. Each label image consists of binary labels representing buildings and non-building (background) areas. For each building polygon k, the contour
C k = { ( x i , y i ) } i = 1 N k
is extracted to define the polygon boundary. From the extracted contour, the centroid c x , k , c y , k and the area A k of the rotated bounding box are calculated as follows:
Polygon   center   point :   c x , k , c y , k = 1 N k i = 1 N k x i , y i ,
Polygon   area :   A k = 1 2 i = 1 N k x i y i + 1 x i + 1 y i .
These geometric attributes of each building polygon are subsequently used during object similarity evaluation and object replacement processes in later stages of the augmentation pipeline.

2.2.2. Multi-Stage Object Replacement Loop

The multi-stage object replacement process replaces building objects in the training dataset with similar objects from other images within the same dataset in a sequential manner. The replacement operation is repeatedly applied to those objects that were not replaced in the first or previous stages. Compared with a single replacement step, performing replacements over multiple stages improves the overall ratio of object replacement and increases the structural diversity of the augmented dataset. From the training dataset, m candidate images are randomly selected, excluding the image that contains the target object to be replaced. Instead of searching through all available images, random selection reduces computational cost and ensures that diverse object shapes are incorporated into the augmentation process. This random sampling strategy also prevents specific images or objects from being repeatedly selected, thereby avoiding potential imbalance within the training data. Although the number of candidate images is limited to m, the number of candidate object polygons contained in those images remains very large. Therefore, only polygons with an area similar to that of the target polygon are selected as candidate polygons. The similarity between the target object polygon and each candidate polygon is measured using the Intersection over Union (IoU) metric. Only candidate object polygons Y that satisfy the following condition are retained for similarity comparison:
A Y A X A X τ a ,
where A Y and A X denote the areas of the candidate polygon Y and the target polygon X , respectively, and τ a is the predefined area similarity threshold.
Each candidate polygon Y is then rotated so that the angle of its rotated bounding box matches the rotation angle θ X of the target polygon X . The similarity between the target polygon X and the rotated candidate polygon Y R is measured using the IoU:
I o U X , Y R = X Y R X Y R ,
where X Y R is the number of pixels common to both X and Y R , and X Y R is the number of pixels contained in their union. The final candidate polygon Y O is selected as the rotated candidate polygon Y R that satisfies the following condition:
Y O = a r g Y R m a x   I o U ( X , Y R ) ,   I o U ( X , Y R ) τ I o U
This replacement approach is not a simple random selection but a precise matching strategy that considers both the size and shape of the candidate polygons. By selecting candidate polygons that closely resemble the target polygon in size and shape, the visual consistency between the replaced polygon and its surrounding background is effectively preserved after the replacement process.
To replace the object polygon X , the polygon X is first removed from the image, and the pixels of the final candidate object polygon Y O , aligned with the rotation angle θ X of the rotated bounding box (RBB), are copied into the position corresponding to polygon X .
To replace a target building, we confine the operation to the target footprint: the candidate building content is clipped by the target polygon mask so that replacement occurs strictly within the original building boundary. This design prevents introducing additional overlaps with surrounding classes (e.g., roads or vegetation) beyond those already present in the source annotation. To reduce boundary seams and abrupt local-context changes, we apply feathering blending along the polygon boundary. Feathering blending is applied to minimize visual discontinuities between the replaced polygon Y O and the surrounding background near the polygon boundary. The feathering process gradually smooths the transition of color and intensity between the replaced polygon Y O and the original background. When the internal pixel intensity of polygon Y O is I Y x , y and the pixel intensity of polygon X in the original image is I X x , y , the synthesized intensity I O x , y of the replaced polygon Y O is determined by the weighting function w x , y , which controls the relative influence of I Y x , y and I X x , y according to the pixel position x , y within the object:
I O x , y = w x , y I Y x , y + 1 w x , y I X x , y ,       M x , y = 255 I X x , y ,                                                                                                                     M x , y = 0         .
To reduce seam artifacts after object insertion, we apply feathered blending along the building boundary. Specifically, a fixed feathering distance of 2 pixels is used to smooth the intensity transition between the replaced building object and the surrounding background while preserving boundary sharpness. Here, the background is defined by the binary building mask (label = 0), and no additional thresholding is introduced beyond the provided mask.
During visual inspection of the augmented samples, we occasionally observed a radiometric seam in bright-background scenes where both roofs and surrounding areas are highly reflective. In a small subset of augmented images, a few replaced roofs appeared slightly darker than their immediate surroundings. This artifact was rare and limited to a small number of buildings in a small portion of the augmented training set, while the building footprint and label geometry remained unchanged. After geometric replacement and blending, the subsequent histogram matching stage further regularizes brightness and color statistics at the image level.

2.2.3. Object-Count-Based Coverage Filtering

The number and density of objects in images where object polygons have been replaced can vary significantly from image to image. Some images contain a large number of object polygons, many of which have been replaced, whereas others include only a few objects and replacements. When the training dataset contains a high proportion of images with a low object-to-background ratio, the influence of background pixels becomes relatively large during loss computation, thereby reducing the contribution of object regions. As a result, the model may fail to adequately learn the shape and boundary characteristics of buildings. To alleviate this imbalance, images are ranked based on the number of building objects they contain, and those with higher object counts are preferentially selected.
After replacing objects, the number of objects in each image is counted and sorted in descending order, and the images with the highest number of objects are subsequently selected. The number of objects in the label image M i is defined as N i , and all boundaries detected through object contour extraction are regarded as individual objects. When the total number of label images is T and the number of objects sorted in descending order is expressed as N 1 N 2 N T , the object-based coverage corresponding to the top K label images is defined as follows:
C o b j e c t K = i = 1 K N i p = 1 T N p ,
where N i denotes the number of objects in the i -th image after sorting in descending order, and N p represents the number of objects in the p -th image in the original (unsorted) order. Images are progressively selected and included in the training dataset until the object-based coverage C object K reaches or exceeds a predefined ratio α .

2.2.4. Object-Based Geometric Augmentation

The object-based augmentation contributes to improving data diversity and domain generalization by replacing building objects in the training images with other building objects of similar size and shape within the same dataset. However, OBA alone has difficulty sufficiently simulating various geometric variabilities, such as changes in orientation, scale and resolution mismatch, and positional misalignment of objects resulting from differences in acquisition angle, resolution, and projection. To enhance robustness to such spatial deformations, we apply geometric augmentation to the images generated through OBA. In this step, we use a standard 2D geometric transformation (Equation (9)) to explicitly introduce controlled variations in orientation, scale, and position. This formulation allows us to simulate spatial deformations commonly observed across domains and to improve robustness to acquisition- and projection-related geometry changes. Moreover, the same transformation is applied to both the image and its corresponding mask/polygon to preserve label–image consistency. The transformations considered are
T T f l i p h o r , T f l i p v e r , T r o t o r t h , T r o t r a n d , T s c a l e , T t r a n s .
Each transformation represents, in order, horizontal flip, vertical flip, orthogonal rotation, random rotation, scaling, and translation, and the definition of each transformation is as follows:
T f l i p h o r x , y = w x , y ,
T f l i p v e r x , y = x , h y ,
T r o t o r t h x , y = R k 90 ° x , y , k 1,2 , 3 ,
T r o t r a n d x , y = R θ x , y , θ U θ m i n , θ m a x ,
T s c a l e x , y = s x , s y , s U s m i n , s m a x ,
T t r a n s x , y = x + t x , y + t y , t x , t y U t m i n , t m a x ,
where R θ is a two-dimensional rotation matrix, and U a , b represents a uniform distribution in the interval a , b , θ m i n and θ m a x represent the rotation angle range, s m i n and s m a x represent the scaling factor range, and t m i n and t m a x represent the translation distance range.
Geometric transformations are applied to both training and label images through a two-step randomization process. In the first step, one of the defined transformations is randomly selected. In the second step, the parameters of the selected transformation (e.g., rotation angle, scaling factor, translation distance) are randomly sampled within a predefined parameter range:
ϕ = k , θ , s , t x , t y .
If the number of OBA-augmented images is N OBA , and geometric transformations are applied k 1 times to each, the total number of geometrically transformed images becomes k 1 · N OBA . The value of k 1 is empirically determined by analyzing changes in model generalization performance as data diversity increases. Through these geometric transformations—such as rotation, scaling, and translation—the spatial variability of the OBA-generated images is further expanded.

2.3. Histogram-Based Normalization and Enhanced Geometric Augmentation

Differences in sensors, acquisition time, and imaging regions often cause mismatches in color distribution between the training and target datasets. Such discrepancies can intensify distributional gaps between domains. In this study, we propose a method that combines histogram matching (HM) with object-based augmentation (OBA) to align the color distribution of training images with that of target images. This approach mitigates domain discrepancies and reduces the model’s performance degradation on the target dataset.
Importantly, the proposed HM is unpaired and non-temporal: it does not require adjacent tiles, overlapping regions, or time-series image pairs. For each training image, the reference (target) image is selected from an unlabeled target-domain pool based solely on distribution similarity, and the transformation is applied globally to the entire image.
During the target candidate selection process, candidate target images are selected based on the histogram of the background regions in the training images where large inter-domain color distribution differences exist. The histogram matching transformation is then applied to the entire image to maintain consistent color tones across the scene. In this study, we use a scenario-specific, fixed number of randomly sampled target candidates (m) for EMD-based matching. The values of m (set to 10% of the training tiles per scenario) are summarized in Appendix A (Table A2). The histogram distribution difference between a fixed number of randomly selected target images and the background of the training image is computed using the Earth Mover’s Distance (EMD) metric. Target images with smaller EMD values are selected as high-ranking candidates. A smaller EMD indicates that the target image has a radiometric distribution closer to that of the training image background. This reduces the risk of unrealistic color shifts and boundary artifacts during subsequent histogram matching, resulting in more stable cross-domain normalization. Among these selected candidates, the target image exhibiting the lowest Shannon entropy—indicating a more uniform and stable color distribution—is chosen as the final target image. By performing histogram matching between this final target image and the training image, the color and luminance distribution of the training image are aligned with the statistical distribution of the target image, thereby correcting color mismatches between domains.
Finally, enhanced geometric augmentation (EGA) is applied to the histogram-normalized images. This step extends the morphological diversity of the training dataset through geometric transformations, thereby enhancing the model’s structural generalization capability. The alignment of color distributions achieved by histogram normalization, combined with the shape and structural diversity introduced through geometric augmentation, contributes to reducing the negative effects of domain gaps on overall model performance.

2.3.1. Histogram Normalization

To compare the distributions between the training and target images, we first define the normalized histogram and the cumulative distribution function (CDF). For channel c of image I , let n c i denote the number of pixels with intensity value i , and let N be the total number of pixels. The normalized histogram h c i and CDF H c i are given by
h I c k = n k c N ,   k = 0 255 h I c k = 1 ,
H I c k = j = 0 k h I c j ,   k 0,255 .
Various distance metrics can be considered when comparing the distributions between the training and target images. The L2 distance metric, while simple to compute, can produce large differences even when there is only a simple shift between distributions [46]. The Kullback–Leibler divergence (KL) can precisely reflect information differences between distributions [47] but becomes unstable in sparse data regions [46]. The Chi-square distance is sensitive to shape differences between distributions [46,48] but is vulnerable to noise [46,49].
In contrast, the Earth Mover’s Distance (EMD) calculates the difference between two distributions from the perspective of transport cost, making it more stable under shifts in luminance or color distributions [46,50]. In this study, EMD is selected as the distance metric because it is robust to inter-domain differences in illumination and color distributions. Let H s c ( k ) and H t c ( k ) denote the CDFs of the training and target images, respectively. Then, the EMD for each channel c is defined as
d E M D c h s c , h t c = k = 0 255 H s c k H t c k .
The overall EMD, obtained by summing the distances across all three color channels, is given by
d E M D h s , h t = c R , G , B d E M D c ( h s c , h t c ) .
During the process of aligning the training image distribution to that of the target image through histogram matching, there is a possibility that the color distribution of the training image may become excessively distorted or overly smoothed, resulting in information loss. In particular, detailed characteristics of pixels in low-contrast regions may be lost, and excessive loss of fine visual information—such as object boundaries and texture—may occur. To address this issue, Shannon entropy is additionally applied as an evaluation metric among the EMD-based target candidates.
The Shannon entropy of an image I is defined as
H I = k = 0 255 h I k l o g ( h I k + ϵ ) ,
where ϵ is a small constant introduced to ensure numerical stability.
If the distribution of pixel intensities is concentrated within a specific range, the entropy value becomes low, indicating a small amount of information. Conversely, if the pixel intensity distribution is evenly spread across the entire range, the entropy value is high, implying a larger amount of information content. The difference Δ H between the entropy of the training image I s and the entropy of the histogram-matched image I s t serves as an indicator of the degree of information preservation in the training image before and after histogram matching, and is expressed as follows:
Δ H = H I s H I s t .
First, for the training image h s and the target image primary candidate set h t j j = 1 M , the EMD between the training image and each target image is computed as:
d j = d E M D ( h s , h t ) , j = 1 , , M .
Based on the EMD values, the top K target images are selected, and a secondary target candidate set C K is formed as follows:
C K = { I t j j argsort ( d j ) [ : K ] } .
Among the images in the secondary candidate set, the target image I t that minimizes the histogram entropy difference Δ H with respect to the training image I s is selected as the final target image I t * :
I t * = arg min I t C K H I s H I s t .
Finally, histogram matching is performed between the final target image I t * and the training image I s to generate the final transformed image.

2.3.2. Enhanced Geometric Augmentation

Histogram matching mitigates illumination- and color-based domain discrepancies between the training and target images. However, this process may reduce color diversity and diminish the visual variations introduced by geometric transformations. To compensate for this, geometric transformations are reapplied after histogram matching to expand color-independent geometric diversity.
The enhanced geometric augmentation randomly selects k 2 · N 0 images—where N 0 is the number of object-based augmented images—from the histogram-normalized image set D norm . For each selected image and its corresponding label, one of the previously defined geometric transformations is randomly chosen and applied. The transformation parameters are also randomly sampled within predefined parameter ranges. The image set generated through enhanced geometric augmentation, denoted as D g e o _ a u g , is then merged with the original image set D o r i g i n a l to form the final training dataset D f i n a l :
D f i n a l = D o r i g i n a l + D g e o _ a u g .
The final training dataset enhances domain transfer adaptability by integrating the structural diversity obtained through histogram matching and geometric augmentation with the luminance and color distribution statistics of the original images.
The outputs and data format are summarized as follows. All intermediate and final outputs are organized in a structured directory hierarchy, including the original images, OBA-augmented images, histogram-normalized images, and the corresponding label masks. Data are stored in standard PNG format with one-to-one filename correspondence between images and masks. This structure is computer-friendly for common deep learning pipelines and user-friendly for visual inspection. Final predictions are exported as binary masks aligned with the input tiles; when needed, the masks can be vectorized into building polygons for GIS use. Key settings required for re-implementation and verification are summarized in Appendix A (Table A3).

3. Experiments and Results

3.1. Experimental Setup and Parameter Configuration

3.1.1. Dataset Partitioning

The OEM dataset was partitioned following a Leave-One-Continent-Out (LOCO) strategy: five of the six continents were used as the training set, while the remaining continent served as the validation and test set. Here, “continent” refers to the OpenEarthMap metadata category used to group scenes by continent, and each continent is treated as a geographic domain. This configuration resulted in six distinct LOCO scenarios, following the OEM benchmark protocol [45].
To avoid spatial leakage, we construct splits at the geographic-unit level rather than by randomly sampling images. For OEM, we adopt a leave-one-continent-out (LOCO) protocol; within the held-out continent, regions are partitioned into disjoint validation and test subsets (approximately half-and-half), and all images inherit the split assignment of their corresponding region. For KOMPSAT-3A, the target city is spatially partitioned into non-overlapping sub-areas that are assigned to validation and test, ensuring geographic separation between the two subsets; the resulting train/validation/test image counts are summarized in Appendix A (Table A4). Due to the large size of the OEM dataset, the training and validation subsets were randomly sampled at 20% of the original dataset size, while the test subset was used in full without reduction. Using these scenario-specific datasets, we analyzed the extent to which the proposed Hybrid (OBA + HM) technique mitigates domain mismatches and improves generalization performance under global conditions characterized by differences in acquisition time, sensor type, and illumination.
For the KOMPSAT-3A dataset, we constructed three cross-city scenarios by designating two of the three cities as training datasets and the remaining city as the validation and test dataset: Seoul + Incheon → Jeju (S + I → J), Jeju + Seoul → Incheon (J + S → I), and Incheon + Jeju → Seoul (I + J → S). In our notation, the cities on the left-hand side form the source domain used for training, whereas the city on the right-hand side is treated as the target domain for validation/testing (e.g., S + I → J trains on Seoul and Incheon and evaluates on Jeju). Using this cross-city setup, we analyze inter-city domain discrepancies driven by brightness, color, and geometric variations. Together with the OEM LOCO setting described above, this yields global (OEM) and regional (KOMPSAT-3A) transfer scenarios for evaluating Hybrid OBA–HM under diverse domain shifts.

3.1.2. Network Training and Implementation

To evaluate model accuracy, all methods were trained and tested with the same segmentation backbone to ensure a fair comparison. We used a U-Net architecture [1] with a VGG-19 encoder [51] initialized from ImageNet weights. Training was performed using the AdamW optimizer (decoupled weight decay) with a learning rate of 1 × 10−4 and a weight decay of 1 × 10−5, a batch size of 16, and 10 epochs. No early stopping was applied; instead, the final model for reporting was selected using the median F1 score over the last three epochs on the validation set. The loss function was binary cross-entropy. All experiments were repeated three times, and both the mean and standard deviation of the performance metrics were reported. All data augmentation was applied offline prior to training (Section 2.2 and Section 2.3), and no additional online augmentation was used during training. All inputs were prepared as 256 × 256 image tiles and min–max scaled to [0, 1] before being fed to the network. Experiments were implemented in Python using TensorFlow–Keras, and all results were produced on a workstation equipped with two NVIDIA RTX A6000 GPUs (48 GB each), an Intel i9-10980XE CPU (3.0 GHz), and 128 GB DDR4-3200 RAM. Model predictions are exported as binary segmentation masks (uint8) aligned with the input tiles, and they can optionally be vectorized to building polygons for GIS use.

3.1.3. Augmentation Parameters

The augmentation parameters of the proposed method consisted of combinations of the number of object replacement iterations k 1 and the number of enhanced geometric augmentation iterations k 2 within the Hybrid OBA–HM pipeline. Performance analysis was conducted for each ( k 1 , k 2 ) combination to determine the optimal configuration, which was then applied consistently in all subsequent experiments.
The number of candidate images for polygon replacement was set to 10% of the total number of training images in each scenario, considering both the replacement ratio and computational efficiency. The area similarity threshold τ A and the IoU-based shape similarity threshold τ IoU were each set to 0.85. The object-count-based coverage parameter η was set to 0.8, and the feathering distance was set to 2 pixels to minimize color distortion within the replaced objects. The number of object replacement stages was determined considering both dataset characteristics and computational efficiency. For the OEM dataset, which includes a wide variety of building types across multiple global regions and is used in multiple domain scenarios, a single-stage replacement was performed. In contrast, for the KOMPSAT-3A dataset, which is relatively smaller and exhibits a higher degree of homogeneity among building objects in the training data, a two-stage replacement was conducted.

3.2. Object-Based Augmentation Results

We analyzed the results of applying the proposed OBA method to the OEM and KOMPSAT-3A datasets. Table 1 presents the results of object-based augmentation on the OEM dataset (single-stage replacement). The proportion of replaced objects across the six scenarios ranged from 24.8% to 30.1%, with an average of 27.65% ± 2.06%. The mean IoU similarity of all replaced polygons across all scenarios was 0.895 ± 0.001, and the average IoU standard deviation was 0.037 ± 0.001. These results indicate that the replaced polygons exhibit a very high degree of shape similarity to the original polygons and that both geometric consistency and positional alignment were stably maintained throughout the augmentation process.
Among the augmented images, the overall average ratio of the final augmented images included within the object-count-based coverage across all scenarios was 32.95% ± 1.28%. This result demonstrates that the augmented images were stably generated at a consistent proportion across all scenarios, indicating the reliability and reproducibility of the augmentation process.
Table 2 shows the object replacement results after performing two-stage object replacement on the KOMPSAT-3A dataset. Between the first and second replacement stages, the augmented image generation rates increased from 70.95% to 73.63% for S + I → J, from 73.17% to 75.42% for J + S → I, and from 42.20% to 45.92% for I + J → S. The average increase in augmented image generation between the first and second stages was 2.88% ± 0.62%, indicating stable growth across all scenarios. The average IoU similarity of all replaced polygons was 0.887 ± 0.006, and the mean IoU standard deviation was 0.041 ± 0.006. The average proportion of object-augmented images across the three scenarios was 27.07% ± 4.37%.
Although the characteristics of the KOMPSAT-3A dataset differ significantly from those of the OEM dataset, the average IoU similarity of replaced polygons in each dataset remained close to 0.89, and the standard deviation was below 0.05. This demonstrates that the proposed object-based augmentation method performs stable polygon replacement even when domain characteristics change. Additionally, while the average augmentation rate in the KOMPSAT-3A dataset was slightly lower and its standard deviation slightly higher than that of the OEM dataset, the deviation remained below 5%, confirming that the proposed augmentation process maintains high stability across domains with differing characteristics.
Figure 3 presents sample results of applying object-based augmentation to the OEM and KOMPSAT-3A datasets. Each example consists of the original image, the augmented image, and the label image. In the label image, replaced buildings are highlighted in green, whereas unreplaced buildings are shown in white. In the OEM example (Figure 3a–c), partial object replacements were applied, and the replaced objects were seamlessly blended into the context of the original scene, thereby preserving a high degree of visual naturalness and contextual coherence. In the KOMPSAT-3A example (Figure 3d–f), most building objects were replaced; however, geometric consistency was successfully maintained. The replaced objects remained spatially coherent with the surrounding roads and adjacent buildings in terms of position, rotation, and orientation, demonstrating natural spatial alignment. This consistency was achieved through rotational alignment and IoU-based candidate filtering during the object replacement process. Consequently, even with OBA applied independently, the proposed method qualitatively demonstrates its capability to achieve visual and structural consistency in both partial replacement scenarios (OEM) and high-coverage replacement scenarios (KOMPSAT-3A).

3.3. Performance on Global and Regional Scenarios

We analyzed the effect of varying the object replacement repetition number k 1 and the enhanced geometric augmentation repetition number k 2 used in the proposed Hybrid (OBA + HM) method on model performance across different domains. The experiments were conducted using the OEM dataset for the global domain and the KOMPSAT-3A dataset for the regional domain environment.
For the OEM dataset, the average accuracy was calculated by performing three trials for each ( k 1 , k 2 ) parameter combination across the six LOCO scenarios, and the overall mean and standard deviation were obtained. For the KOMPSAT-3A dataset, two of the three cities were used for training and the remaining city for validation and testing; the average accuracy from three trials was calculated for each parameter combination, and the mean and standard deviation across the three cross-city scenarios were computed.
Accuracy evaluation for each parameter combination was conducted by first selecting the optimal model based on the training and validation datasets, and then applying the selected model to the test dataset to obtain the final accuracy. The average F1-score values for each parameter combination on both datasets are summarized in Figure 4. Figure 4a shows the results for the OEM dataset, and Figure 4b illustrates the results for the KOMPSAT-3A dataset.

3.3.1. OEM-Scenario (Global Domains)

In the OEM scenario heatmap in Figure 4a, the vertical axis represents the number of object replacement iterations k 1 , the horizontal axis represents the number of geometric augmentation iterations k 2 , and the color scale indicates the F1-score. Overall, the heatmap shows a trend in which the color intensity increases toward the right and bottom, indicating that performance improves as both k 1 and k 2 increase.
However, in the region where k 1 = 8 , the accuracy either stagnated or slightly decreased in some scenarios (1, 2, and 4), suggesting that excessive object replacement may cause distributional distortion. In the horizontal direction k 2 , performance consistently improved from k 2 = 1 to k 2 = 8 ; however, when k 2 4 , the color change became more gradual, indicating that the effect of geometric augmentation had reached saturation. The best performance was achieved at ( k 1 , k 2 ) = ( 4 , 8 ) , with an average F1-score of 0.848 and a standard deviation of σ = 0.0386. Although ( k 1 , k 2 ) = ( 2 , 8 ) and ( k 1 , k 2 ) = ( 4 , 8 ) are both displayed as 0.848 in the heat maps due to rounding to three decimal places, their averaged F1-scores are 0.8477 and 0.8481, respectively, indicating that ( k 1 , k 2 ) = ( 4 , 8 ) is the true maximum configuration. These results demonstrate that ( k 1 , k 2 ) = ( 4 , 8 ) , a balanced combination of object replacement and geometric augmentation, is the most effective configuration for achieving global domain generalization.

3.3.2. KOMPSAT-3A Scenario (Regional Transfer Domains)

In the KOMPSAT-3A scenario heatmap in Figure 4b, the color intensity also becomes progressively darker toward the right and bottom. This trend, similar to that observed in the OEM scenario, indicates that performance improves as both the number of object replacement iterations and geometric augmentation iterations increase. At the bottom-right region of the heatmap, corresponding to ( k 1 , k 2 ) = ( 8 , 8 ) , the average F1-score reached its maximum value of 0.6895. However, the improvement compared with the optimal combination derived from the OEM scenario, ( 4 , 8 ) , was minimal, with a difference of ΔF1 ≈ 0.008. Statistical testing confirmed that the difference in F1-score between the ( 4 , 8 ) and ( 8 , 8 ) combinations was not statistically significant (p = 0.62, α = 0.05). The parameter combination ( k 1 , k 2 ) = ( 4 , 8 ) , identified from the OEM scenario, was thus verified to maintain statistically equivalent performance in both global and regional domain environments. Accordingly, this configuration was adopted as the optimal parameter setting for all subsequent experiments.

3.4. Comparative Evaluation and Statistical Robustness

To quantitatively evaluate the performance of the proposed Hybrid (OBA + HM) method, we compared it with six widely used data augmentation methods. The compared methods include RGT (Random Geometric Transform), RIC (Random Image Copy), PA (Photometric Adjustment), GHM (Global Histogram Matching), CycleGAN, and ColorMapGAN. The main characteristics of these methods and the domain aspects they target are summarized in Table 3. Each augmentation method was independently analyzed using the same datasets and identical model parameter settings to assess performance variations attributable to the augmentation strategies.
Each comparison method was designed to address different types of domain discrepancies. RGT enhances geometric invariance by expanding diversity in scale, orientation, and spatial arrangement. RIC mitigates the imbalance between object and background classes by increasing the number of training samples through random duplication. PA improves robustness to radiometric variations such as lighting and weather conditions by randomly adjusting brightness, hue, and saturation. GHM alleviates cross-domain color mismatch by aligning the color distribution of training images to that of the target domain. CycleGAN transforms source images into the style of the target domain without requiring paired training data, thereby reducing global appearance discrepancies in color tone, texture, and illumination. ColorMapGAN employs GAN-based color mapping to align the color distribution of source images with that of the target domain, effectively mitigating color and radiometric mismatches between domains.
The proposed Hybrid (OBA + HM) method simultaneously achieves geometric diversity through object-based augmentation and geometric transformations, while enhancing robustness to radiometric variations through histogram normalization, thereby providing balanced domain adaptation performance that effectively compensates for both geometric and radiometric discrepancies.
The key parameters for each method were configured as follows: RGT applied rotations of ±30°, scaling between 0.8 and 1.2, translations of ±20 pixels, and a horizontal flip probability of 50%. RIC used random replication of original images with replacement allowed. PA applied random adjustments to brightness (±51), contrast (0.8–1.2), gamma (0.9–1.1), hue (±5°), and saturation (±15%). GHM performed color alignment in Lab color space using weighting factors (0.3/0.2/0.2 for L/a/b) and clipping in the 1–99% range.

3.4.1. OEM Dataset

Table 4 presents the average and standard deviation of the F1-score for each augmentation method, as well as the performance improvement compared with the baseline, using the OEM dataset under global-domain conditions. Here, Baseline denotes training on the original dataset without object-based augmentation (OBA), histogram matching (HM), or enhanced geometric augmentation (EGA), using only the standard preprocessing applied to all methods (e.g., resizing and normalization). Unless otherwise stated, this Baseline definition applies to the comparisons reported in Table 4 and Table 5, and all other training and evaluation settings are kept identical across methods to ensure a fair comparison.
To address the effect of combining OBA with histogram matching, we additionally report OBA-only results and compare them with the proposed Hybrid (OBA + HM). The goal of this comparison is to quantify the incremental, pipeline-level benefit of adding histogram matching after object replacement, rather than to equalize the number of generated samples across pipelines. We note that OBA-only yields fewer unique augmented samples than Hybrid by design because it excludes the HM-based stage and the subsequent augmentation steps. Therefore, the OBA-only vs. Hybrid gap should be interpreted as a practical estimate of the benefit of the combined method under our two-stage framework. All methods are trained and evaluated using identical architectures, hyperparameters, and protocols.
The baseline model achieved an average F1-score of 0.808 ± 0.034. The RGT method, which randomly applied geometric transformations such as rotation, scaling, and translation, yielded a performance improvement of ΔF1 = +0.013 compared with the baseline. This improvement is attributed to the model’s enhanced ability to adapt to variations in building shape, spatial arrangement, and orientation through exposure to diverse geometric transformations. The RIC method increased the dataset size by randomly sampling images from the training dataset, resulting in a modest improvement of ΔF1 = +0.008. This indicates that, in large-scale and multi-domain environments, the effect of duplication was relatively limited, and the increase in dataset volume contributed to a modest enhancement in the stochastic stability of model training. The PA method, which applied random adjustments to brightness, contrast, gamma, hue, and saturation, achieved an improvement of ΔF1 = +0.010, enhancing robustness to variations in color and illumination conditions. The GHM method, which reduces radiometric differences through global histogram matching, showed only a limited improvement of ΔF1 = +0.004, likely because global normalization does not sufficiently capture localized radiometric characteristics present across different domains. CycleGAN provided no overall performance improvement compared with the baseline (ΔF1 ≈ −0.020), and ColorMapGAN achieved only a small improvement of ΔF1 ≈ +0.004. These results suggest that, while both methods contribute to global style and color alignment, they may not sufficiently compensate for structural domain differences and boundary information—such as building size, density, and background composition—which are important in OEM scenarios.
The proposed object-based augmentation alone (OBA-Only) achieved an average F1-score of 0.827 ± 0.031, corresponding to a +0.019 improvement over the baseline. This indicates that object-level replacement enhances structural diversity and improves generalization under global-domain conditions. Compared with OBA-Only, the proposed Hybrid (OBA + HM) further increased the F1-score by 0.013 (0.840 vs. 0.827), suggesting an additional benefit from histogram-based radiometric normalization when combined with object-level augmentation.
In contrast, the proposed Hybrid (OBA + HM) method achieved an average F1-score of 0.840 ± 0.026, corresponding to a +0.032 improvement over the baseline. This superior performance is attributed to the combination of object-level structural diversity and histogram-based radiometric normalization, which jointly mitigates both geometric and radiometric mismatches. Moreover, the standard deviation (±0.026) was similar to or lower than those of the other methods, indicating that the proposed approach yields statistically stable and reproducible results.

3.4.2. KOMPSAT-3A Dataset

Table 5 presents the results of the cross-city domain transfer experiments conducted using the KOMPSAT-3A dataset. The baseline model achieved an average F1-score of 0.455 ± 0.198, exhibiting substantial performance instability caused by structural variations across cities—including differences in building density, orientation, and architectural shape. The RGT method achieved a performance improvement of ΔF1 = +0.043, indicating that geometric transformations effectively corrected inter-city geometric discrepancies and improved overall accuracy. The RIC method resulted in a performance decline of ΔF1 = −0.012, suggesting that, in small-scale and relatively homogeneous domains, simply increasing the dataset size does not enhance structural diversity and may instead lead to overfitting to specific city patterns. The PA method yielded the largest performance gain among the single augmentation methods, achieving ΔF1 = +0.046. This finding implies that, despite the use of the same sensor, differences in illumination and color still exist between domains, and compensating for radiometric diversity is effective in improving transfer performance. The GHM method resulted in a performance decrease of ΔF1 = −0.019, indicating that global histogram normalization can locally reduce image contrast, thereby degrading accuracy. CycleGAN produced a performance degradation of ΔF1 = −0.023 compared with the baseline, whereas ColorMapGAN showed a relatively small improvement of ΔF1 = +0.043. These results suggest that, while both CycleGAN and ColorMapGAN methods contribute to reducing differences in illumination and color between cities, they have limitations in independently addressing the structural domain gaps prominent in the KOMPSAT-3A dataset, such as building density, background patterns, and geometric arrangements.
The proposed object-based augmentation alone (OBA-Only) improved the baseline to 0.573 ± 0.174 (ΔF1 = +0.118), indicating that object-level replacement and geometric augmentation effectively increase structural diversity across cities. Compared with OBA-Only, the proposed Hybrid (OBA + HM) further improved performance by 0.079 (0.652 vs. 0.573), highlighting the additional benefit of histogram-based radiometric normalization when combined with object-level augmentation.
Overall, the proposed Hybrid (OBA + HM) method achieved a significant performance improvement of ΔF1 = +0.197 over the baseline, with an average F1-score of 0.652 ± 0.119. This substantial gain can be attributed to the combined effects of object-level structural diversification through object replacement and geometric transformations, together with enhanced radiometric consistency achieved via histogram normalization. These complementary mechanisms effectively mitigate inter-city discrepancies in shape, illumination, and radiometric conditions.
To summarize the quantitative results across both OEM (Table 4) and KOMPSAT-3A (Table 5) scenarios, we additionally compare Hybrid OBA–HM against a pre-defined strongest baseline. Because each transfer scenario includes only three runs (n = 3), per-scenario paired tests are underpowered; thus, we primarily interpret effect sizes and cross-scenario consistency. We pre-define the strongest baseline as the non-hybrid method with the highest overall mean F1 across all evaluated scenarios (Random Geometric TR in our experiments; Appendix A, Table A5). Hybrid OBA–HM improves over the strongest baseline in all nine scenarios, supported by a two-sided sign test (p = 0.0039) and a one-sided Wilcoxon signed-rank test (p = 0.00195) on the scenario-level mean differences Δ(H − Strongest) (Appendix A, Table A5).
Figure 5 presents a qualitative comparison of building segmentation results for each augmentation method on a sample image from the OEM dataset. Each visualization includes the original image, ground truth, and the results obtained from the baseline (no augmentation), RGT, RIC, PA, GHM, CycleGAN, ColorMapGAN, and the proposed Hybrid (OBA + HM) method. In the segmentation maps, white represents true positives (TP), black represents true negatives (TN), red indicates false negatives (FN), and green denotes false positives (FP).
The baseline results showed the simultaneous occurrence of numerous FN and FP pixels, resulting in unstable detection of object boundaries and fine-scale structural details. The RGT method achieved the lowest FN occurrence through diverse geometric transformations, but produced the largest number of FP pixels, leading to noticeable over-segmentation. The RIC, PA, and GHM methods yielded a moderate level of FN occurrence, while FP pixels decreased compared with RGT but remained non-negligible. The GHM method produced slightly more FN pixels than the Hybrid method, while the FP pixels appeared at a nearly similar level. CycleGAN augmentation resulted in relatively few FP pixels but produced the largest number of FN pixels among all comparison methods. ColorMapGAN yielded segmentation performance comparable to that of GHM, but localized FP pixels were still observed in some regions. In contrast, the proposed Hybrid (OBA + HM) technique yielded the fewest FN pixels and, together with PA, the lowest level of FP pixels, qualitatively confirming its geometric and radiometric correction effects.
Figure 6 shows a qualitative comparison of object-detection results for each augmentation technique using a sample image from the KOMPSAT-3A dataset. Compared with the baseline, the RGT method resulted in increased FN occurrence, and the RIC, PA, and GHM techniques also exhibited slightly larger FN pixels relative to the baseline. The CycleGAN and ColorMapGAN methods yielded the largest number of FN pixels among all the approaches. However, the proposed Hybrid (OBA + HM) method exhibited only a small number of FN pixels, with only a few FP pixels observed.
Notably, several objects that were undetected (FN) by all other methods were successfully detected by the Hybrid approach, visually demonstrating the complementary effectiveness of object-based augmentation and histogram normalization. These results indicate that the Hybrid method maintains stable and consistent object-detection performance even in the KOMPSAT-3A environment characterized by significant inter-city domain discrepancies, thereby demonstrating its superior cross-domain generalization capability.

3.5. Domain-Wise Indicator and Heatmap Analysis

In this section, we quantitatively analyze which domain indicators influence the performance improvement of the proposed method. By combining this quantitative analysis with heatmap visualization, we compare and interpret the relative characteristics of domain mismatch factors observed in the global (OEM) and regional (KOMPSAT-3A) domains. In Section 3.5.1, the correlations between the proposed domain indicators—brightness difference (ΔMean V), color distribution ( EMD ( HS ) ), building variance Var(HS)_Building, and directional mismatch ( EMD ( θ ) )—and the ΔF1 improvements are analyzed to evaluate the performance sensitivity to each domain discrepancy factor. Then, in Section 3.5.2, we visually compare and interpret the types of mismatches (radiometric vs. geometric) between the OEM and KOMPSAT-3A domains using normalized domain-indicator heatmaps. This comprehensive analysis demonstrates that the proposed Hybrid (OBA + HM) method effectively mitigates complex domain mismatches by simultaneously compensating for both radiometric and geometric discrepancies.

3.5.1. ΔF1 vs. Domain Indicators

To quantitatively analyze how the proposed Hybrid (OBA + HM) method outperforms Photometric Adjustment (PA) across different domain dimensions such as brightness, color, and structure, the ΔF1 (=Hybrid − PA) improvement was compared with four key domain indicators: brightness difference (ΔMean V), color distribution distance (EMD(HS)), variance of building chromaticity (Var(HS)_Building), and directional mismatch (EMD(θ)). The relationships between each domain indicator and ΔF1 are illustrated in Figure 7.
ΔF1 vs. ΔMean V (Brightness Bias): As the inter-domain brightness variation (ΔMean V) increased, ΔF1 tended to increase, indicating a positive association. This pattern is consistent with histogram matching enhancing radiometric stability compared with PA by mitigating inter-domain brightness discrepancies through global histogram normalization. For both the OEM and KOMPSAT-3A datasets, the performance gain in ΔF1 increased with greater brightness differences, suggesting that the proposed Hybrid (OBA + HM) method helps mitigate radiometric discrepancies between domains caused by brightness variations.
ΔF1 vs. EMD(HS) (Color Distribution Distance): As the color and saturation distribution difference EMD(HS) increased, ΔF1 showed an increasing trend, indicating a positive association. This trend is consistent with histogram matching reducing color bias by normalizing the mean and variance of the color distribution in the training domain toward the statistical distribution of the target domain. Furthermore, OBA may contribute additional performance gains over PA by enhancing intra-object color diversity, resulting in improved robustness to color and radiometric variations across domains.
ΔF1 vs. Var(HS)_Building (Chroma Diversity): Regarding building chromatic diversity, ΔF1 remained stable across varying levels of chromatic variation, as shown in Figure 7c, indicating no clear association or fluctuation between the two variables. This suggests that the proposed method maintains positive performance gains even when chromatic diversity varies. This stability is consistent with the OBA process increasing the structural diversity of objects and improving model robustness under diverse geometric and illumination conditions. Additionally, the HM stage performs global chromatic normalization, stabilizing the mean and variance of the color distribution and minimizing color variability across domains.
ΔF1 vs. EMD(θ) (Orientation Distribution Shift): As the directional distribution difference EMD(θ) increased, ΔF1 tended to increase, as illustrated in Figure 7d, indicating a positive association. This trend is consistent with the combination of two-stage geometric enhancement k 1 , k 2 ) = ( 4 , 8 and object replacement (OBA) being beneficial under larger orientation shifts. In particular, for the KOMPSAT-3A dataset, in inter-city transfer cases such as I + J→S, where EMD(θ) values were large, ΔF1 exceeded +0.25, showing a substantial performance improvement over PA. These observations suggest that the proposed Hybrid (OBA + HM) method can maintain strong performance gains even in domains exhibiting significant geometric discrepancies.
As the number of transfer scenarios is limited, we retained all points in Figure 7 and interpreted high-variance cases as informative rather than removing them. As a robustness check, we computed Spearman’s rank correlation between Δ F 1 and each domain indicator. The rank-based associations remained positive for Δ Mean V ( ρ = 0.650 ) and EMD(HS) ( ρ = 0.683 ), whereas Var(HS)_Building showed no monotonic relationship ( ρ = 0.117 ).
Across all four domain indicators, ΔF1 exhibited positive or statistically stable performance improvements, and the proposed method consistently outperformed PA across the dimensions of brightness, color, and geometry. In particular, the largest positive slopes were observed along the EMD(HS) and EMD(θ) axes, suggesting that the relative performance advantage of the proposed method becomes more pronounced as color and directional mismatches increase. Consequently, the Hybrid (OBA + HM) method achieved improved cross-domain generalization performance over PA across diverse domain factors by combining global brightness normalization and two-stage geometric enhancement k 1 , k 2 ) = ( 4 , 8 . This trend is further supported in the normalized domain-indicator analysis presented in Section 3.5.2, which suggests that the configuration k 1 , k 2 ) = ( 4 , 8 is a robust and broadly applicable parameter setting within the evaluated OEM and KOMPSAT-3A scenarios. We emphasize that these trends indicate correlation rather than causation; since the indicators may covary, a more rigorous multivariate analysis (e.g., partial correlation or regression with collinearity control) is left for future work.

3.5.2. Normalized Domain-Indicator Heatmap

Figure 8 shows the comparison results of the domain indicators—ΔMean V, ΔStd V, EMD(HS), Var(HS)_Building, EMD(log A), and EMD(θ)—normalized to a 0–100 range for six OEM (global) scenarios and three KOMPSAT-3A (regional) scenarios. The purpose of this analysis is to quantitatively compare the relative discrepancies of brightness, color, and directional factors across each dataset and scenario, thereby identifying the dominant discrepancy patterns of each domain group.
The normalization results showed that the OEM scenarios exhibited low values for most indicators, ranging between 0 and 25%, indicating a homogeneous domain with minimal brightness, color, and orientation discrepancies. In contrast, the KOMPSAT-3A scenarios showed values of ΔMean V, ΔStd V, EMD(HS), Var(HS)_Building, and EMD(θ) mostly ranging from 30 to 100%. Several indicators exhibited high deviations exceeding 60%, indicating a high-discrepancy domain characterized by significantly greater composite mismatches than the OEM domain. These results suggest that the OEM dataset represents a globally radiometrically consistent and structurally homogeneous environment, whereas the KOMPSAT-3A dataset represents a more complex environment with combined radiometric and geometric variations arising from urban structural diversity and differences in imaging conditions.
In the OEM scenarios, ΔMean V (0–16.9%), ΔStd V (0–24.4%), EMD(HS) (0–21.5%), and EMD(θ) (0–21.6%) all showed low levels, indicating an overall homogeneous domain with minimal inter-regional differences in illumination, color, and directionality. Var(HS)_Building showed an intermediate level (0–69%), and EMD(log A) also exhibited intermediate values; however, in some scenarios (e.g., S5, S6), relatively high values around 90% (89–100%) were observed, indicating partial deviations in building area distributions. These findings demonstrate that the OEM domain is generally a low-discrepancy environment with minimal radiometric and geometric mismatches, although certain scenarios show regional variations in building area distributions. Furthermore, the orientation mismatch EMD(θ) was nearly absent, indicating that building orientation was consistently maintained across domains. Accordingly, the HM stage of the Hybrid (OBA + HM) method contributes to maintaining radiometric consistency within the OEM domain by correcting subtle brightness and color deviations rather than applying extreme normalization.
In the KOMPSAT-3A scenarios, significantly greater deviations were observed across all major indicators compared with the OEM scenarios. ΔMean V (30.3–100%) and ΔStd V (33.7–100%) showed substantial brightness and illumination differences between cities, while EMD(HS) (24.8–100%) revealed strong radiometric mismatches due to substantial variation in color distribution range and centroid. Additionally, Var(HS)_Building (33–100%) and EMD(θ) (2.5–100%) indicated simultaneous increase in the internal chromatic diversity of buildings and mismatches in building orientations, demonstrating that radiometric and geometric factors act in a strongly coupled manner across the regional domain. To mitigate these compound mismatches, the proposed Hybrid (OBA + HM) method reduced structural and orientation discrepancies between cities by combining OBA for structural diversification and two-stage geometric enhancement ( k 1 , k 2 ) = ( 4 , 8 ) . Furthermore, the HM stage globally normalized brightness and color distributions, thereby mitigating radiometric mismatches between domains and complementarily reinforcing the effects of both the OBA and geometric enhancement stages. As a result, within the KOMPSAT-3A domain, the proposed Hybrid (OBA + HM) method achieved a significantly greater radiometric and geometric adaptability gain than in the OEM domain. The combined effects of brightness and color normalization (HM) and object-based structural and directional correction (OBA and geometric enhancement) enabled stable and consistent cross-domain generalization performance, even under highly complex radiometric–geometric discrepancy conditions.
The OEM (Global) domain can be characterized as a radiometric-dominant domain, exhibiting low deviations in brightness, color, and directionality, where the HM stage effectively contributed to global brightness and color normalization. In contrast, the KOMPSAT-3A (Regional) domain can be classified as a radiometric–geometric composite domain, showing substantially greater deviations not only in brightness and color but also in internal chromatic diversity and orientation mismatches compared with the OEM domain. In this domain, OBA combined with multi-stage geometric enhancement ( k 1 = 4 , k 2 = 8 ) effectively alleviated structural and directional mismatches, while HM compensated for radiometric biases through global brightness and color normalization. Consequently, the proposed Hybrid (OBA + HM) method was identified as a globally robust configuration that adaptively responds to both radiometric and geometric factors, achieving stable and consistent cross-domain generalization performance across both global and regional domains.

4. Discussion

4.1. The Role and Implications of Object-Based Augmentation

Object-based augmentation (OBA) plays a key role in ensuring structural diversity of buildings by generating diverse building instances from limited training datasets. In the OEM and KOMPSAT-3A datasets, the mean IoU between replaced and original building polygons is approximately 0.895 and 0.887, respectively, with a standard deviation below 0.05. This suggests that the size, shape, and location of the replaced objects are well aligned with those of the original buildings. This suggests that, unlike simple random replication-based augmentation, OBA increases structural diversity while maintaining object-level geometric consistency. In addition, OBA uses object-count-based coverage filtering to select augmented images with a high proportion of building objects, which alleviates the bias toward the background class during loss computation and increases the effective contribution of the building class during training.
In the KOMPSAT-3A inter-city transfer experiments, where building density and occupancy rates vary significantly, the Hybrid OBA–HM scheme substantially improved the average F1-score compared with the baseline while simultaneously reducing its standard deviation. This suggests more stable performance in our experiments even in environments with pronounced inter-domain structural variations in density, shape, and layout. In summary, OBA functions as a data-level module that compensates for structural domain differences without altering the network architecture, thereby directly contributing to improved domain generalization performance within the evaluated settings.

4.2. Hybrid Response to Radiometric-Geometric Complex Mismatch

Comparative experiments show that the Hybrid OBA–HM approach, which combines object-based structural augmentation with histogram-based normalization, achieves larger performance gains than augmentations that address only a single type of domain factor, such as geometric transformation (RGT) or radiometric adjustment (PA, GHM). This suggests that the two stages operate in a complementary manner. In the OEM global scenario, global histogram matching (GHM) produced only a marginal improvement over the baseline (ΔF1 ≈ +0.004), whereas the Hybrid OBA–HM approach achieved ΔF1 ≈ +0.032. This result indicates that even in domains such as OEM, where overall radiometric and geometric mismatches are relatively small, the domain gap can be reduced more effectively when object-level structural augmentation and global radiometric regularization are applied together. This tendency was even more evident in the KOMPSAT-3A inter-city transfer experiments. Despite using the same sensor, the KOMPSAT-3A dataset represents a high-discrepancy domain with large inter-city variations in illumination, color, and orientation. In this setting, GHM alone actually degraded performance (ΔF1 ≈ −0.019), indicating that applying only global histogram normalization can lead to a loss of local contrast or edge information. By contrast, the Hybrid OBA–HM approach improved the F1-score from 0.455 to 0.652 (ΔF1 ≈ +0.197), suggesting that domain-gap mitigation is most effective when the two stages operate jointly in environments with complex radiometric and geometric mismatches in our experiments.
These results indicate that, within the proposed framework, HM acts not as a standalone domain adaptation module but rather as an auxiliary normalization stage that aligns brightness and color distributions with the target domain once sufficient structural diversity has been secured by OBA. In other words, OBA primarily handles the structural component, whereas HM controls the style component. Working together, the two stages form a hybrid scheme that jointly expands and aligns the structure–style combination space learned by the model within the evaluated settings.

4.3. Domain Metric-Based Interpretation and Practical Implications

Domain-indicator analysis quantitatively clarifies which domain factors primarily explain the performance gains of Hybrid OBA–HM. Analysis of the relationships between ΔF1 (Hybrid − PA) and differences in brightness, color distribution, and directionality distribution shows that larger discrepancies in all three metrics are associated with greater relative performance improvements of Hybrid OBA–HM. In contrast, no significant correlation was observed between the building color diversity index and ΔF1, indicating that the performance gain of the proposed technique remains relatively robust even the chromatic diversity within buildings varies. This behavior can be interpreted as the result of the HM stage normalizing global mean and variance of the color distribution, thereby absorbing inter-domain differences in color diversity, while OBA compensates for object structure and thus maintains structural recognition capability under high color complexity. The normalized domain-indicator heatmaps clearly show that the OEM (global) domain is a low-discrepancy environment with small radiometric and geometric deviations overall, whereas the KOMPSAT-3A (regional) domain is a high-discrepancy environment with large combined radiometric–geometric mismatches. This provides practical implications for data-centric design: domain indicators can be computed in advance to diagnose which axes (e.g., brightness, color, orientation) exhibit large mismatches in a given target domain, and augmentation and normalization strategies can then be selected accordingly.
From a practical perspective, the proposed Hybrid (OBA + HM) module has the advantage that it can be integrated into the data preprocessing and augmentation pipeline without modifying the existing semantic segmentation network architecture. In particular, a base model can be trained on multi-domain source data such as OpenEarthMap, and Hybrid OBA–HM can then be applied to imagery from new sensors or cities. In this way, a meaningful level of cross-domain generalization improvement can be achieved within similar conditions to those evaluated here, without deploying complex adversarial networks or target pseudo-label generation.
Building on these practical implications, Hybrid OBA–HM can be combined with model-centric domain adaptation as a data-level preprocessing step. Applying OBA–HM prior to pseudo-labeling/self-training may improve pseudo-label reliability under large shifts, and it may also be paired with feature-space alignment to address residual discrepancies. In such combinations, diminishing returns may occur once major radiometric and structural gaps are reduced; thus, integration should be tuned based on target-domain diagnostics (e.g., brightness/color mismatch indicators), noting that overly aggressive histogram normalization may reduce local contrast in some cases.

4.4. Limitations and Future Work

This study has several limitations. First, because the HM stage is based on global histogram matching, it may not precisely compensate for fine-scale local illumination and material variations within urban scenes. Second, our experiments focused on binary building segmentation; therefore, additional validation is needed for multi-class land cover segmentation and more complex scene compositions. Third, the evaluation was limited to two datasets and a single backbone; thus, the conclusions should be interpreted as evidence of domain generalization gains within these experimental settings rather than as universal claims across all sensors and regions. Nonetheless, Hybrid (OBA + HM) remains attractive in that it (i) can be integrated into existing semantic segmentation pipelines at the data preprocessing and augmentation stages without modifying the model architecture; (ii) improves cross-domain performance using only source-domain labels, without requiring complex adversarial networks or target pseudo-labeling; and (iii) when paired with domain-indicator analysis, provides practical guidance for diagnosing domain gaps and designing augmentation strategies for new domains.
In future research, the proposed technique should be extended to a broader range of remote sensing applications, such as multi-class land cover segmentation, pre- and post-disaster time-series analysis, change detection, and object detection. Furthermore, combining Hybrid OBA–HM with pseudo-label-based self-training and feature-space domain adaptation may lead to a more comprehensive, data-centric domain generalization framework that jointly considers both model and data aspects.

5. Conclusions

In this study, we proposed a hybrid, data-centric framework (Hybrid (OBA + HM)) that integrates Object-Based Augmentation (OBA) and Histogram Matching (HM) to alleviate the domain gap arising from differences in region, sensor, and imaging conditions in remote sensing-based building segmentation. The proposed approach expands structural diversity through object-level replacement and geometric augmentation using label polygons, while mitigating radiometric and geometric deviations between training and target domains through histogram-based normalization and enhanced geometric augmentation with EMD- and Shannon entropy-based target selection.
Across both the OEM global scenarios and the KOMPSAT-3A regional scenarios, Hybrid (OBA + HM) improved cross-domain building segmentation performance over representative augmentation baselines and the non-augmented setting in our experiments. Importantly, the object replacement process maintained high geometric consistency, indicating that the augmented samples preserve realistic building shapes and placements under domain shifts. In addition, the parameter analysis identified a practical iteration setting that provided stable, near-optimal performance across the evaluated scenarios, and we adopted this setting as the default in our experiments.
The domain-indicator-based analysis further provided an interpretable view of when the proposed framework is most beneficial. In particular, performance gains tended to increase as mismatches in brightness, color distribution, and directionality became larger, suggesting that the hybrid design is especially effective under combined radiometric–geometric shifts. The indicator heatmaps also highlighted a contrast between OEM as a relatively low-discrepancy, radiometric-dominant domain and KOMPSAT-3A as a composite mismatch domain, supporting the practical use of domain indicators for diagnosing domain gaps and selecting data-centric augmentation strategies within the evaluated settings.

Author Contributions

Conceptualization, methodology, software, validation, formal analysis, writing—original draft, writing—review and editing, project administration and funding acquisition, C.Y.; data curation, Y.A. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the Korea Agency for Infrastructure Technology Advancement (KAIA) grant funded by the Ministry of Land, Infrastructure and Transport (Grant RS-2022-00155763).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The OpenEarthMap (OEM) dataset used in this study is publicly available at https://zenodo.org/records/7223446 (accessed on 1 December 2025). The KOMPSAT-3A dataset contains high-resolution satellite imagery over the Korean Peninsula. Due to national security regulations and licensing restrictions in the Republic of Korea, the KOMPSAT-3A dataset is not publicly available and cannot be shared. The source code implementing the proposed Hybrid OBA–HM pipeline was developed under an ongoing national R&D project and is subject to institutional and legal restrictions; therefore, it is not publicly available. Due to the same administrative and legal constraints, we are unable to provide the code for individual distribution at this time. To support reproducibility, the manuscript provides detailed descriptions of the algorithms, parameter settings, and training/evaluation protocols, enabling independent re-implementation and verification on publicly available data (e.g., OEM).

Acknowledgments

During the preparation of this manuscript, the authors used ChatGPT 5.1 (OpenAI) for the purposes of assisting in the generation of heatmap and domain-wise indicator visualizations. The authors have reviewed and edited the output and take full responsibility for the content of this publication.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. Supplementary Tables

This appendix provides supplementary tables to support clarity and reproducibility. Table A1 summarizes the notation used in the proposed pipeline. Table A2 lists the number of randomly sampled target candidates (m) used for histogram matching in each scenario. Table A3 reports the dataset splits for OEM and KOMPSAT-3A. Table A4 provides a reproducibility checklist with key implementation settings. Table A5 presents extended scenario-wise F1-scores (mean ± SD over three runs) and effect sizes.
Table A1. Notation used in the proposed pipeline.
Table A1. Notation used in the proposed pipeline.
Symbol/TermDescription
D o r i g i n a l Original image set (images and labels)
D n r o m Histogram-normalized image set
D g e o _ a u g Image set generated through enhanced geometric augmentation
D f i n a l Final training dataset
D norm Histogram-normalized image set
X Target building polygon to be replaced
Y Candidate building polygon selected for replacement
Y R Rotated candidate polygon
Y O Final candidate polygon content after alignment
RBBRotated bounding box used for polygon alignment
θ X Rotation angle of the RBB for polygon X
I X x , y Pixel intensity inside polygon X at location x , y
I Y x , y Pixel intensity inside candidate polygon Y O at location x , y
I O x , y Synthesized pixel intensity after replacement/blending
w x , y Feathering weight controlling the blend ratio at x , y
IoUIntersection-over-Union used for polygon similarity/quality checking
EMDEarth Mover’s Distance used as a histogram/domain-distance metric
k 1 Number of object-replacement rounds
k 2 Number of geometric-augmentation rounds
Table A2. Number of randomly sampled target candidates (m) used for histogram matching (10% of training tiles per scenario).
Table A2. Number of randomly sampled target candidates (m) used for histogram matching (10% of training tiles per scenario).
ScenarioTraining Imagesm (10%)
OEM-S16874688
OEM-S25441544
OEM-S36080608
OEM-S47452745
OEM-S56548655
OEM-S67435744
S + I→J7491750
J + S→I6906691
I + J→S 8235824
Table A3. Reproducibility checklist: key settings for re-implementation and verification.
Table A3. Reproducibility checklist: key settings for re-implementation and verification.
CategoryEssential ItemSpecification (This Study)
DataDatasets and availabilityOEM (public; see Data Availability Statement); KOMPSAT-3A (restricted; not publicly shareable due to national security/licensing constraints).
DataOEM split protocolLOCO (6 scenarios); geographic-unit (region)-based splitting to avoid spatial leakage; within held-out continent, regions split into disjoint validation/test (~50/50); train/val randomly sampled at 20%, test at 100%.
DataKOMPSAT-3A split protocol3 cross-city scenarios: S + I→J, J + S→I, I + J→S; target city partitioned into non-overlapping validation/test sub-areas; image counts in Appendix A (Table A4).
PreprocessInput formation256 × 256 tiles; min–max scaling to [0, 1]; RGB channels.
ModelArchitectureU-Net with VGG-19 encoder initialized from ImageNet weights.
TrainingCore hyperparametersAdamW (lr 1 × 10−4; weight decay 1 × 10−5); batch size 16; 10 epochs; binary cross-entropy loss; no early stopping.
TrainingReporting protocol3 runs per experiment; mean ± standard deviation reported; model selected by median validation F1 over last three epochs.
AugmentationApplication modeOffline only (Section 2.2 and Section 2.3); no online augmentation during training.
Hybrid
OBA–HM
Key parameters(k1, k2) grid evaluated; candidate pool = 10% of training images/scenario; τ_A = 0.85; τ_IoU = 0.85; η = 0.8; feathering = 2 pixels; stages: OEM 1-stage, KOMPSAT-3A 2-stage.
InferenceBinarization &
post-processing
Threshold = 0.4; no morphological post-processing (opening/closing, small-component removal).
Outputs/EvalOutput and metricsBinary masks (uint8) aligned with tiles; standard segmentation metrics including F1-score computed consistently across methods; optional polygon vectorization for GIS use.
SoftwareRuntime environmentPython 3.10.0; TensorFlow 2.8.2; Keras 2.10.0; implementation in TensorFlow–Keras.
SystemPlatformWindows 11 Pro; CUDA 11.2; cuDNN 8 (per TensorFlow build information); 2× NVIDIA RTX A6000 (48 GB each); Intel i9-10980XE (3.0 GHz); 128 GB RAM.
ReproducibilityRandomness statementRandom seeds were not fixed globally; each experiment was repeated three times and results are reported as mean ± standard deviation.
DataDatasets and availabilityOEM (public; see Data Availability Statement); KOMPSAT-3A (restricted; not publicly shareable due to national security/licensing constraints).
Table A4. Train/validation/test image counts used in OEM and KOMPSAT-3A experiments.
Table A4. Train/validation/test image counts used in OEM and KOMPSAT-3A experiments.
ScenarioTraining ImagesValidation ImagesTest Images
OEM-S16874552541
OEM-S2544112551271
OEM-S360801008878
OEM-S47452252263
OEM-S56548685733
OEM-S67435218314
S + I → J749123502444
J + S → I690627652765
I + J → S 823518721911
Table A5. Scenario-wise F1-scores (mean ± SD over three runs) for the Original baseline, the strongest baseline, and Hybrid OBA–HM, with effect sizes.
Table A5. Scenario-wise F1-scores (mean ± SD over three runs) for the Original baseline, the strongest baseline, and Hybrid OBA–HM, with effect sizes.
ScenarioOriginal (Mean ± SD)Strongest Baseline (Mean ± SD)Hybrid OBA–HM (Mean ± SD)Δ(H − Strongest)Δ(H − Original)
OEM-S10.790 ± 0.0400.809 ± 0.0230.833 ± 0.0030.0250.044
OEM-S20.801 ± 0.0050.809 ± 0.0070.823 ± 0.0100.0140.022
OEM-S30.794 ± 0.0140.807 ± 0.0050.813 ± 0.0060.0060.02
OEM-S40.787 ± 0.0440.817 ± 0.0490.828 ± 0.0040.0110.041
OEM-S50.821 ± 0.0220.837 ± 0.0260.853 ± 0.0050.0150.032
OEM-S60.857 ± 0.0140.848 ± 0.0490.887 ± 0.0040.0390.03
S + I → J0.668 ± 0.0370.705 ± 0.0230.744 ± 0.0120.0380.076
J + S → I0.550 ± 0.0180.523 ± 0.0910.698 ± 0.0430.1750.149
I + J → S0.147 ± 0.0530.265 ± 0.1810.515 ± 0.1030.250.368
Notes: Values are mean ± standard deviation over three runs. Δ(H − Strongest) and Δ(H − Original) denote mean F1 differences. The strongest baseline is pre-defined a priori as the non-hybrid method with the highest overall mean F1 across all evaluated scenarios (Random Geometric TR in this study).

References

  1. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015; Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F., Eds.; Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2015; Volume 9351, pp. 234–241. [Google Scholar] [CrossRef]
  2. Teo, T.; Chen, P. Building change detection in aerial imagery using end-to-end deep learning semantic segmentation techniques. Buildings 2025, 15, 695. [Google Scholar] [CrossRef]
  3. Yang, S.; Peng, B.; Sui, T.; Wu, M.; Huang, Q. Advancing self-supervised learning for building change detection and damage assessment. Remote Sens. 2025, 17, 2717. [Google Scholar] [CrossRef]
  4. Ansari, R.A.; Malhotra, R.; Ansari, M.Z. Building footprint identification using remotely sensed images: A compressed sensing-based approach to support map updating. Geomatics 2025, 5, 7. [Google Scholar] [CrossRef]
  5. Li, W.; He, C.; Fang, J.; Zheng, J.; Fu, H.; Yu, L. Semantic segmentation-based building footprint extraction using very high-resolution satellite images and multi-source GIS data. Remote Sens. 2019, 11, 403. [Google Scholar] [CrossRef]
  6. Milosavljević, A. Automated processing of remote sensing imagery using deep semantic segmentation: A building footprint extraction case. ISPRS Int. J. Geo-Inf. 2020, 9, 486. [Google Scholar] [CrossRef]
  7. Al Shafian, S.; He, C.; Hu, D. DamageScope: An integrated pipeline for building damage segmentation, geospatial mapping, and interactive web-based visualization. Remote Sens. 2025, 17, 2267. [Google Scholar] [CrossRef]
  8. Yang, Y.; Bian, W.; Fang, J.; Tang, M.; He, Z.; Li, Y.; Fan, G. Application of MambaBDA for building damage assessment in the 2025 Los Angeles wildfire. Buildings 2025, 15, 4019. [Google Scholar] [CrossRef]
  9. Liu, J.; Luo, Y.; Chen, S.; Wu, J.; Wang, Y. BDHE-Net: A novel building damage heterogeneity enhancement network for accurate and efficient post-earthquake assessment using aerial and remote sensing data. Appl. Sci. 2024, 14, 3964. [Google Scholar] [CrossRef]
  10. Zhao, F.; Zhang, C.; Zhang, R.; Wang, T. Visual prompt learning of foundation models for post-disaster damage evaluation. Remote Sens. 2025, 17, 1664. [Google Scholar] [CrossRef]
  11. Tilon, S.; Nex, F.; Kerle, N.; Vosselman, G. Post-disaster building damage detection from earth observation imagery using unsupervised and transferable anomaly detecting generative adversarial networks. Remote Sens. 2020, 12, 4193. [Google Scholar] [CrossRef]
  12. Gu, J.; Xie, Z.; Zhang, J.; He, X. Advances in rapid damage identification methods for post-disaster regional buildings based on remote sensing images: A survey. Buildings 2024, 14, 898. [Google Scholar] [CrossRef]
  13. Hacıefendioğlu, K.; Başağa, H.B.; Kahya, V.; Özgan, K.; Altunışık, A.C. Automatic detection of collapsed buildings after the 6 February 2023 Türkiye earthquakes using post-disaster satellite imagery with deep learning-based semantic segmentation models. Buildings 2024, 14, 582. [Google Scholar] [CrossRef]
  14. Ji, Y.; Wu, W.; Wan, G.; Zhao, Y.; Wang, W.; Yin, H.; Tian, Z.; Liu, S. Segment Anything Model-based building footprint extraction for residential complex spatial assessment using LiDAR data and very high-resolution imagery. Remote Sens. 2024, 16, 2661. [Google Scholar] [CrossRef]
  15. Han, Z.; Li, X.; Wang, X.; Wu, Z.; Liu, J. Building segmentation in urban and rural areas with MFA-Net: A multidimensional feature adjustment approach. Sensors 2025, 25, 2589. [Google Scholar] [CrossRef]
  16. Yu, A.; Quan, Y.; Yu, R.; Guo, W.; Wang, X.; Hong, D.; Zhang, H.; Chen, J.; Hu, Q.; He, P. Deep learning methods for semantic segmentation in remote sensing with small data: A survey. Remote Sens. 2023, 15, 4987. [Google Scholar] [CrossRef]
  17. Lu, X.; Gong, T.; Zheng, X. Multisource compensation network for remote sensing cross-domain scene classification. IEEE Trans. Geosci. Remote Sens. 2020, 58, 2504–2515. [Google Scholar] [CrossRef]
  18. Banerjee, B.; Bovolo, F.; Bhattacharya, A.; Bruzzone, L.; Chaudhuri, S.; Buddhiraju, K.M. A novel graph-matching-based approach for domain adaptation in classification of remote sensing image pair. IEEE Trans. Geosci. Remote Sens. 2015, 53, 4045–4062. [Google Scholar] [CrossRef]
  19. Tsai, Y.-H.; Hung, W.-C.; Schulter, S.; Sohn, K.; Yang, M.-H.; Chandraker, M. Learning to adapt structured output space for semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 7472–7481. [Google Scholar] [CrossRef]
  20. Chen, Y.; Li, W.; Sakaridis, C.; Dai, D.; Van Gool, L. No more discrimination: Cross city adaptation of road scene segmenters. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2011–2020. [Google Scholar] [CrossRef]
  21. Deng, X.; Zhu, Y.; Tian, Y.; Newsam, S. Scale-aware adaptation for land-cover classification in remote sensing imagery. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 5–9 January 2021; pp. 2160–2169. [Google Scholar] [CrossRef]
  22. Zeng, J.; Gu, Y.; Qin, C.; Jia, X.; Deng, S.; Xu, J.; Tian, H. Unsupervised domain adaptation for remote sensing semantic segmentation with the 2D discrete wavelet transform. Sci. Rep. 2024, 14, 23552. [Google Scholar] [CrossRef]
  23. Benjdira, B.; Bazi, Y.; Koubaa, A.; Ouni, K. Unsupervised domain adaptation using generative adversarial networks for semantic segmentation of aerial images. Remote Sens. 2019, 11, 1369. [Google Scholar] [CrossRef]
  24. Benjdira, B.; Ammar, A.; Koubaa, A.; Ouni, K. Data-efficient domain adaptation for semantic segmentation of aerial imagery using generative adversarial networks. Appl. Sci. 2020, 10, 1092. [Google Scholar] [CrossRef]
  25. Tasar, O.; Happy, S.L.; Tarabalka, Y.; Alliez, P. ColorMapGAN: Unsupervised domain adaptation for semantic segmentation using color mapping generative adversarial networks. IEEE Trans. Geosci. Remote Sens. 2020, 58, 7178–7193. [Google Scholar] [CrossRef]
  26. Mateo-García, G.; Laparra, V.; López-Puigdollers, D.; Gómez-Chova, L. Cross-sensor adversarial domain adaptation of Landsat-8 and Proba-V images for cloud detection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 747–761. [Google Scholar] [CrossRef]
  27. Gao, K.; Yu, A.; You, X.; Guo, W.; Li, K.; Huang, N. Integrating multiple sources knowledge for class asymmetry domain adaptation segmentation of remote sensing images. IEEE Trans. Geosci. Remote Sens. 2024, 62, 1–18. [Google Scholar] [CrossRef]
  28. Li, Z.; Xie, Y.; Jia, X.; Stuart, K.; Delaire, C.; Skakun, S. Point-to-region co-learning for poverty mapping at high resolution using satellite imagery. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), Washington, DC, USA, 7–14 February 2023; pp. 14321–14328. [Google Scholar] [CrossRef]
  29. Iqbal, J.; Ali, M. Weakly supervised domain adaptation for built-up region segmentation in aerial and satellite imagery. ISPRS J. Photogramm. Remote Sens. 2020, 167, 263–275. [Google Scholar] [CrossRef]
  30. Zou, Y.; Yu, Z.; Kumar, B.V.K.V.; Wang, J. Unsupervised domain adaptation for semantic segmentation via class-balanced self-training. In Computer Vision—ECCV 2018; Springer: Cham, Switzerland, 2018; pp. 289–305. [Google Scholar] [CrossRef]
  31. Pham, H.; Dai, Z.; Xie, Q.; Le, Q.V. Meta pseudo labels. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 11557–11568. [Google Scholar] [CrossRef]
  32. Cermelli, F.; Mancini, M.; Rota Bulò, S.; Ricci, E.; Caputo, B. Modeling the background for incremental and weakly-supervised semantic segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 10099–10113. [Google Scholar] [CrossRef]
  33. Saha, S.; Zhao, S.; Zhu, X.X. Multitarget domain adaptation for remote sensing classification using graph neural network. IEEE Geosci. Remote Sens. Lett. 2022, 19, 6506505. [Google Scholar] [CrossRef]
  34. Zhu, X.X.; Tuia, D.; Mou, L.; Xia, G.-S.; Zhang, L.; Xu, F.; Fraundorfer, F. Deep learning in remote sensing: A comprehensive review and list of resources. IEEE Geosci. Remote Sens. Mag. 2017, 5, 8–36. [Google Scholar] [CrossRef]
  35. Illarionova, S.; Nesteruk, S.; Shadrin, D.; Ignatiev, V.; Pukalchik, M.; Oseledets, I. Object-Based Augmentation Improves Quality of Remote Sensing Semantic Segmentation. arXiv 2021, arXiv:2105.05516. [Google Scholar] [CrossRef]
  36. Illarionova, S.; Nesteruk, S.; Shadrin, D.; Ignatiev, V.; Pukalchik, M.; Oseledets, I. Object-Based Augmentation for Building Semantic Segmentation: Ventura and Santa Rosa Case Study. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV) Workshops, Montreal, BC, Canada, 11–17 October 2021; pp. 1659–1668. [Google Scholar] [CrossRef]
  37. Motoi, I.M.; Saraceni, L.; Nardi, D.; Ciarfuglia, T.A. Evaluating the Efficacy of Cut-and-Paste Data Augmentation in Semantic Segmentation for Satellite Imagery. In Proceedings of the IGARSS 2024—2024 IEEE International Geoscience and Remote Sensing Symposium, Athens, Greece, 7–12 July 2024. [Google Scholar] [CrossRef]
  38. Zhang, C.; Yue, P.; Cigna, F.; Tapete, D. OCA: Object-Based Change Augmentation for Few-Shot Building Change Detection in Very High-Resolution Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2025, in press. [Google Scholar] [CrossRef]
  39. Chen, H.; Li, W.; Shi, Z. Adversarial Instance Augmentation for Building Change Detection in Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5603216. [Google Scholar] [CrossRef]
  40. Wangiyana, S.; Samczyński, P.; Gromek, A. Data Augmentation for Building Footprint Segmentation in SAR Images: An Empirical Study. Remote Sens. 2022, 14, 2012. [Google Scholar] [CrossRef]
  41. Yaras, C.; Kassaw, B.A.; Bradbury, K.; Malof, J.M. Randomized Histogram Matching: A Simple Augmentation for Unsupervised Domain Adaptation in Overhead Imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 1988–1998. [Google Scholar] [CrossRef]
  42. Ma, J. Histogram Matching Augmentation for Domain Adaptation with Application to Multi-Centre, Multi-Vendor and Multi-Disease Cardiac Image Segmentation. In Statistical Atlases and Computational Models of the Heart. M&Ms and EMIDEC Challenges; Bernard, O., Camara, O., Zhuang, X., Eds.; Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2021; Volume 12592, pp. 177–186. [Google Scholar] [CrossRef]
  43. Ma, H.; Lin, X.; Wu, Z.; Yu, Y. Coarse-to-Fine Domain Adaptive Semantic Segmentation with Photometric Alignment and Category-Center Regularization. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Virtual, 19–25 June 2021; pp. 4050–4060. [Google Scholar] [CrossRef]
  44. Ghosh Mondal, T.; Shi, Z.; Zhang, H.; Chen, G. Class-Wise Histogram Matching-Based Domain Adaptation in Deep Learning-Based Bridge Element Segmentation. J. Civ. Struct. Health Monit. 2025, 15, 1973–1989. [Google Scholar] [CrossRef]
  45. Xia, J.; Yokoya, N.; Adriano, B.; Broni-Bediako, C. OpenEarthMap: A Benchmark Dataset for Global High-Resolution Land Cover Mapping. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 2–7 January 2023; pp. 6254–6264. [Google Scholar] [CrossRef]
  46. Rner, Y.; Puzicha, J.; Tomasi, C.; Buhmann, J.M. Empirical evaluation of dissimilarity measures for color and texture. Comput. Vis. Image Underst. 2001, 84, 25–43. [Google Scholar] [CrossRef]
  47. Kullback, S.; Leibler, R.A. On information and sufficiency. Ann. Math. Stat. 1951, 22, 79–86. [Google Scholar] [CrossRef]
  48. Asha, V.; Bhajantri, N.U.; Nagabhushan, P. GLCM-based chi-square histogram distance for automatic detection of defects on patterned textures. Int. J. Comput. Vis. Robot. 2011, 2, 302–313. [Google Scholar] [CrossRef]
  49. Pele, O.; Werman, M. The quadratic-chi histogram distance family. In Computer Vision—ECCV 2010; Daniilidis, K., Maragos, P., Paragios, N., Eds.; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2010; Volume 6312, pp. 749–762. [Google Scholar] [CrossRef]
  50. Rubner, Y.; Tomasi, C.; Guibas, L.J. The earth mover’s distance as a metric for image retrieval. Int. J. Comput. Vis. 2000, 40, 99–121. [Google Scholar] [CrossRef]
  51. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar] [CrossRef]
Figure 1. Flowchart of the proposed Hybrid OBA–HM method, from object replacement and normalization to model training and evaluation.
Figure 1. Flowchart of the proposed Hybrid OBA–HM method, from object replacement and normalization to model training and evaluation.
Applsci 16 00543 g001
Figure 2. Representative examples of input images from the OEM and KOMPSAT-3A datasets. (a) A European scene from the OEM dataset, where buildings are mostly isolated; (bd) KOMPSAT-3A images over Seoul, Incheon, and Jeju, respectively, where buildings are closely packed and the overall building layout exhibits a pronounced directional pattern.
Figure 2. Representative examples of input images from the OEM and KOMPSAT-3A datasets. (a) A European scene from the OEM dataset, where buildings are mostly isolated; (bd) KOMPSAT-3A images over Seoul, Incheon, and Jeju, respectively, where buildings are closely packed and the overall building layout exhibits a pronounced directional pattern.
Applsci 16 00543 g002
Figure 3. OBA results for the OpenEarthMap (OEM) and KOMPSAT-3A datasets. Images (ac) represent OEM samples, and images (df) represent KOMPSAT-3A samples. (a,d) show the original scenes, (b,e) show the augmented results, and (c,f) show the label masks. Green indicates replaced buildings, while white denotes non-replaced buildings.
Figure 3. OBA results for the OpenEarthMap (OEM) and KOMPSAT-3A datasets. Images (ac) represent OEM samples, and images (df) represent KOMPSAT-3A samples. (a,d) show the original scenes, (b,e) show the augmented results, and (c,f) show the label masks. Green indicates replaced buildings, while white denotes non-replaced buildings.
Applsci 16 00543 g003
Figure 4. Heat maps of the mean F1-score for different combinations of the parameters ( k 1 , k 2 ) . (a) Results on the OEM dataset. (b) Results on the KOMPSAT-3A dataset. The color of each cell represents the average F1-score obtained for the corresponding ( k 1 , k 2 ) pair.
Figure 4. Heat maps of the mean F1-score for different combinations of the parameters ( k 1 , k 2 ) . (a) Results on the OEM dataset. (b) Results on the KOMPSAT-3A dataset. The color of each cell represents the average F1-score obtained for the corresponding ( k 1 , k 2 ) pair.
Applsci 16 00543 g004
Figure 5. Qualitative comparison of building segmentation results with different data augmentation strategies on the OEM dataset: (a) original image; (b) ground truth; (c) baseline (no augmentation); (d) RGT; (e) RIC; (f) PA; (g) GHM; (h) CycleGAN; (i) ColorMapGAN; (j) proposed Hybrid (OBA + HM).
Figure 5. Qualitative comparison of building segmentation results with different data augmentation strategies on the OEM dataset: (a) original image; (b) ground truth; (c) baseline (no augmentation); (d) RGT; (e) RIC; (f) PA; (g) GHM; (h) CycleGAN; (i) ColorMapGAN; (j) proposed Hybrid (OBA + HM).
Applsci 16 00543 g005
Figure 6. Qualitative comparison of building segmentation results with different data augmentation strategies on the KOMPSAT-3A dataset: (a) original image; (b) ground truth; (c) baseline (no augmentation); (d) RGT; (e) RIC; (f) PA; (g) GHM; (h) CycleGAN; (i) ColorMapGAN; (j) proposed Hybrid (OBA + HM).
Figure 6. Qualitative comparison of building segmentation results with different data augmentation strategies on the KOMPSAT-3A dataset: (a) original image; (b) ground truth; (c) baseline (no augmentation); (d) RGT; (e) RIC; (f) PA; (g) GHM; (h) CycleGAN; (i) ColorMapGAN; (j) proposed Hybrid (OBA + HM).
Applsci 16 00543 g006
Figure 7. Relationship between the relative F1-score gain of the proposed Hybrid (OBA + HM) method over Photometric Adjustment (PA), ΔF1 (=Hybrid − PA), and four domain indicators: brightness difference (ΔMean V), color distribution distance (EMD(HS)), variance of building chromaticity (Var(HS)_Building), and directional inconsistency (EMD(θ)). Dashed lines indicate linear fits; the fitted equation is shown for panels (a,b,d), while panel (c) reports the mean level (y ≈ mean(ΔF1)).
Figure 7. Relationship between the relative F1-score gain of the proposed Hybrid (OBA + HM) method over Photometric Adjustment (PA), ΔF1 (=Hybrid − PA), and four domain indicators: brightness difference (ΔMean V), color distribution distance (EMD(HS)), variance of building chromaticity (Var(HS)_Building), and directional inconsistency (EMD(θ)). Dashed lines indicate linear fits; the fitted equation is shown for panels (a,b,d), while panel (c) reports the mean level (y ≈ mean(ΔF1)).
Applsci 16 00543 g007
Figure 8. Normalized domain indicators (0–100) for six OEM and three KOMPSAT-3A cross-city scenarios: ΔMean V, ΔStd V, EMD(HS), Var(HS)_Building, EMD(Log A), and EMD(θ).
Figure 8. Normalized domain indicators (0–100) for six OEM and three KOMPSAT-3A cross-city scenarios: ΔMean V, ΔStd V, EMD(HS), Var(HS)_Building, EMD(Log A), and EMD(θ).
Applsci 16 00543 g008
Table 1. Object replacement results on the OEM dataset (Single-Stage Replacement).
Table 1. Object replacement results on the OEM dataset (Single-Stage Replacement).
ScenarioTraining ImagesAugmented ImagesAugmentation Ratio (%)Total ObjectsReplaced ObjectsReplacement Ratio (%)Mean IoUIoU Std. Dev
OEM-S16874231433.744,26210,95524.80.8930.036
OEM-S25441191835.344,45912,38727.90.8960.037
OEM-S36080197632.537,942949625.00.8940.037
OEM-S47452243832.752,36915,04628.70.8950.036
OEM-S56548204531.249,13614,80330.10.8960.037
OEM-S67435240332.352,50615,42829.40.8960.037
Table 2. Object replacement results on the KOMPSAT-3A dataset (Two-Stage Replacement).
Table 2. Object replacement results on the KOMPSAT-3A dataset (Two-Stage Replacement).
ScenarioTraining ImagesAugmented ImagesAugmentation Ratio (%)Total ObjectsReplaced ObjectsReplacement Ratio (%)Mean IoUIoU Std. Dev
S + I → J7491156220.981,17759,77273.60.8820.045
J + S → I6906210230.487,76666,19775.40.8830.044
I + J → S 8235246529.956,14725,78345.90.8950.034
Table 3. Data augmentation methods used in this study, their key characteristics, and targeted domain aspects.
Table 3. Data augmentation methods used in this study, their key characteristics, and targeted domain aspects.
MethodFull NameKey CharacteristicsTargeted Domain Aspect
RGTRandom Geometric TransformRotation, scaling, translation, and other geometric transformationsGeometric diversity (shape, layout, orientation)
RICRandom Image CopyRandom sampling with replacement from original imagesQuantity-driven diversity (increased sample volume)
PAPhotometric AdjustmentRandom adjustment of brightness, contrast, gamma, hue, and saturationRadiometric diversity
GHMGlobal Histogram MatchingRandom target selection and global histogram matchingDomain color alignment
CycleGANCycle-Consistent Generative Adversarial NetworkUnpaired image-to-image translation with cycle-consistency constraints between source and target domainsBidirectional appearance adaptation between domains
ColorMapGANColor Mapping Generative Adversarial NetworkGAN-based pixelwise color mapping that aligns the spectral distribution of source images to that of target images while preserving scene structureSpectral/radiometric distribution alignment between domains
OBA-OnlyObject-Based Augmentation (OBA) OnlyPolygon-guided building instance replacement with geometric alignment and basic geometric transforms (no histogram matching)Geometric/structural diversity (building shape, density, layout)
Hybrid (OBA + HM) (Proposed Method)Object-Based Augmentation + Histogram MatchingObject-level augmentation with histogram-based domain normalizationCombined geometric and radiometric adaptation
Table 4. Performance comparison of data augmentation strategies on the OEM dataset. The table reports the mean F1-score, standard deviation, and ΔF1 relative to the baseline.
Table 4. Performance comparison of data augmentation strategies on the OEM dataset. The table reports the mean F1-score, standard deviation, and ΔF1 relative to the baseline.
MethodMean F1 ± Std. DevΔF1 (Compared with Baseline)
Baseline0.808 ± 0.034
RGT0.821 ± 0.0310.013
RIC0.816 ± 0.0270.008
PA0.818 ± 0.0250.010
GHM0.812 ± 0.0320.004
CycleGAN0.788 ± 0.040−0.020
ColorMapGAN0.812 ± 0.0270.004
OBA-Only0.827 ± 0.0310.019
Hybrid (OBA + HM)0.840 ± 0.0260.032
Note: OBA-only yields fewer augmented samples than Hybrid by design; thus, the comparison is reported as a pipeline-level incremental benefit.
Table 5. Performance comparison of data augmentation strategies on the KOMPSAT-3A dataset. The table reports the mean F1-score, standard deviation, and ΔF1 relative to the baseline.
Table 5. Performance comparison of data augmentation strategies on the KOMPSAT-3A dataset. The table reports the mean F1-score, standard deviation, and ΔF1 relative to the baseline.
MethodMean F1 ± Std. DevΔF1 (Compared with Baseline)
Baseline0.455 ± 0.198
RGT0.498 ± 0.2170.043
RIC0.443 ± 0.267−0.012
PA0.501 ± 0.1910.046
GHM0.436 ± 0.215−0.019
CycleGAN0.432 ± 0.231−0.023
ColorMapGAN0.498 ± 0.1720.043
OBA-Only0.573 ± 0.1740.118
Hybrid (OBA + HM)0.652 ± 0.1190.197
Note: OBA-only yields fewer augmented samples than Hybrid by design; thus, the comparison is reported as a pipeline-level incremental benefit.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ye, C.; Ahn, Y. Hybrid Object-Based Augmentation and Histogram Matching for Cross-Domain Building Segmentation in Remote Sensing. Appl. Sci. 2026, 16, 543. https://doi.org/10.3390/app16010543

AMA Style

Ye C, Ahn Y. Hybrid Object-Based Augmentation and Histogram Matching for Cross-Domain Building Segmentation in Remote Sensing. Applied Sciences. 2026; 16(1):543. https://doi.org/10.3390/app16010543

Chicago/Turabian Style

Ye, Chulsoo, and Youngman Ahn. 2026. "Hybrid Object-Based Augmentation and Histogram Matching for Cross-Domain Building Segmentation in Remote Sensing" Applied Sciences 16, no. 1: 543. https://doi.org/10.3390/app16010543

APA Style

Ye, C., & Ahn, Y. (2026). Hybrid Object-Based Augmentation and Histogram Matching for Cross-Domain Building Segmentation in Remote Sensing. Applied Sciences, 16(1), 543. https://doi.org/10.3390/app16010543

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop