Next Article in Journal
Development of Variable Elastic Band with Adjustable Elasticities for Semi-Passive Exosuits
Previous Article in Journal
Quantitative Assessment of CFD-Based Micro-Scale Renovation of Existing Building Component Envelopes
Previous Article in Special Issue
Bio-Inspired Multi-Granularity Model for Rice Pests and Diseases Named Entity Recognition in Chinese
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Water Body Identification from Satellite Images Using a Hybrid Evolutionary Algorithm-Optimized U-Net Framework

1
School of Computer Science and Technology, Chongqing University of Posts and Telecommunications, Chongqing 400065, China
2
Chongqing Institute of Green and Intelligent Technology, Chinese Academy of Sciences, Chongqing 400714, China
3
School of Software Engineering, Chengdu University of Information Technology, Chengdu 610225, China
4
Dazhou Key Laboratory of Government Data Security, Sichuan University of Arts and Science, Dazhou 635000, China
5
Chongqing Institute of Meteorological Sciences, Chongqing 401147, China
6
Xinjiang Technical Institute of Physics & Chemistry, Chinese Academy of Sciences, Urumqi 830011, China
*
Authors to whom correspondence should be addressed.
Biomimetics 2025, 10(11), 732; https://doi.org/10.3390/biomimetics10110732
Submission received: 9 September 2025 / Revised: 22 October 2025 / Accepted: 29 October 2025 / Published: 1 November 2025

Abstract

Accurate and automated identification of water bodies from satellite imagery is critical for environmental monitoring, water resource management, and disaster response. Current deep learning approaches, however, suffer from a strong dependence on manual hyperparameter tuning, which limits their automation capability and robustness in complex, multi-scale scenarios. To overcome this limitation, this study proposes a fully automated segmentation framework that synergistically integrates an enhanced U-Net model with a novel hybrid evolutionary optimization strategy. Extensive experiments on public Kaggle and Sentinel-2 datasets demonstrate the superior performance of our method, which achieves a Pixel Accuracy of 96.79% and an F1-Score of 94.75, outperforming various mainstream baseline models by over 10% in key metrics. The framework effectively addresses the class imbalance problem and enhances feature representation without human intervention. This work provides a viable and efficient path toward fully automated remote sensing image analysis, with significant potential for application in large-scale water resource monitoring, dynamic environmental assessment, and emergency disaster management.

1. Introduction

High-precision water body extraction holds significant practical value for global water coverage dynamic monitoring, water resource storage assessment, urban water environment management, disaster emergency response, and the protection of aquatic ecologically sensitive areas [1,2]. Particularly in the current context of intensifying global climate change and given the increasing frequency of extreme hydrological events, the development of efficient and reliable water segmentation techniques has become a pressing need. In recent years, driven by breakthrough advances in computer vision and artificial intelligence technologies, water segmentation techniques have undergone a significant technological generational leap [3].
Early research on water body identification and extraction primarily relied on physically prior-driven threshold models [4]. These included rapid segmentation techniques utilizing near-infrared band absorption characteristics [5] or enhanced water indices (e.g., MNDWI, AWEI) [6,7], alongside methods like edge detection operators [8], data mining [9], and object-based approaches [10]. In recent years, deep learning has achieved a technological leap forward through autonomous feature learning and hierarchical modeling [11]. The Fully Convolutional Network (FCN) stands as a milestone in CNN-based semantic segmentation methods [12]. U-Net [13], VGG [14], MobileNet [15], ResNet-50 [16], and Densenet [17] have been extensively applied to water body identification tasks due to their robust feature extraction capabilities and end-to-end learning paradigm. For instance, in 2020, Wang et al. [17] proposed a multidimensional densely connected convolutional neural network for water body identification, which further enhanced model performance. Additionally, Franz et al. [18] systematically compared 32 convolutional neural network architectures, demonstrating that the ResNet50-encoder-based U-Net model achieved optimal performance on their self-constructed river water segmentation dataset.
While the aforementioned models primarily focus on mitigating the impact of spectral reflectance variability (influenced by factors such as weather conditions, solar illumination angles, and land cover types) and noise interference (e.g., clouds, shadows, buildings) to enhance water body identification accuracy, and have achieved certain improvements in precision, their performance remains highly dependent on hyperparameter selection and network architecture design. Manual parameter tuning is not only time-consuming and labor-intensive but also often fails to locate the global optimum, making it difficult to strike an optimal balance between accuracy and computational efficiency. Consequently, model robustness is constrained.
Training deep learning models requires pre-setting certain hyperparameters, such as the network architecture, activation function, and learning rate [19]. These hyperparameter settings directly determine the performance and practical application of deep learning models. Currently, there is no definitive formula for selecting hyperparameters; instead, they rely primarily on a combination of experimentation and experience. This has become a major factor influencing the successful application of deep learning models across various problems. For deep learning models, hyperparameters include weights and hyperparameters. While optimization methods like Adam [20] offer good results for weights, their results are significantly affected by hyperparameters. Improper hyperparameter settings can prevent convergence during model training, leading to suboptimal results. Therefore, efficient hyperparameter optimization algorithms are crucial for deep learning models implemented in remote sensing image interpretation. Our main contributions can be summarized as follows:
(1)
An evolutionary-hybrid neural framework for automated optimization of water body segmentation models. We develop a novel approach that integrates a multi-algorithm evolutionary optimization system (combining GA, ACO, PSO, and AE) with deep neural networks to automate hyperparameter configuration for water body identification. This framework effectively addresses the critical challenge of manual parameter tuning in deep learning-based water segmentation methods by dynamically optimizing the learning rate, batch size, and momentum, thereby significantly enhancing model performance and training efficiency while reducing human intervention.
(2)
Adopting a learnable multi-loss function fusion strategy with adaptive weighting, a linearly weighted loss function fusion strategy is proposed to collaboratively optimize classification confidence and geometric consistency. Through adaptive adjustments during training, the contributions of different loss terms to model optimization are balanced, addressing the class imbalance problem in water segmentation. Weights are automatically adjusted during training using an evolutionary algorithm to ensure an optimal balance between the various loss function objectives throughout the learning process.
(3)
The method was verified on two public remote sensing datasets. It significantly outperformed the baseline model in multiple quantitative indicators (such as mIoU and F1_Score), especially when dealing with complex remote sensing scenes such as urban areas (building shadow interference), cloud cover (noise interference), and multi-scale water bodies (from small rivers to large lakes). It demonstrated stronger generalization ability and boundary segmentation accuracy, providing a better automated solution for the automated high-precision extraction of remote sensing water bodies.

2. Related Work

In this section, first, we first review related work from two perspectives: deep learning-based water body interpretation methods and deep model parameter optimization using evolutionary strategies.

2.1. Water Body Interpretation Method Based on Deep Learning Model

Deep learning primarily utilizes artificial neural networks to autonomously learn and extract features from remote sensing imagery, thereby avoiding potential human bias caused by manual feature selection and reducing reliance on prior knowledge. Deep learning demonstrates high accuracy and robustness when processing remote sensing imagery with complex backgrounds and diverse textures [21]. For example, Chen Qian et al. [22] used the DeepLabv3 semantic segmentation model to isolate water features from high-resolution satellite imagery, mitigating the effects of shadows and buildings more effectively than traditional classification and machine learning techniques. The fully convolutional neural network (FCN) and its variants, such as UNet, SegNet, and DeepLab, have become the mainstream deep learning semantic segmentation models for water extraction and are widely used in water extraction from remote sensing imagery [23,24]. The integration of deep learning and remote sensing technology has made new avenues for surface water extraction, water environment monitoring, and the study of global hydrological phenomena. The above methods have achieved remote sensing water body segmentation to a certain extent, but most of these methods require manual adjustment of classifier hyperparameters.

2.2. Deep Model Parameter Optimization

Evolutionary Algorithms (EAs) [25,26], which are global optimization methods based on the principles of biological evolution, have been introduced to optimize deep learning models. By simulating biological processes such as natural selection, crossover, and mutation, these evolutionary algorithms can efficiently search for optimal solutions in complex parameter spaces. Common evolutionary algorithms include the Genetic Algorithm (GA) [27], Ant Colony Optimization (ACO) [28,29], and Particle Swarm Optimization (PSO) [30]. These frameworks achieve high-precision performance while focusing on automated model optimization, efficient architecture discovery, and robust training strategies, significantly reducing reliance on manual intervention. Li et al. [31] introduced a new evolutionary computation framework aided by machine learning, named EVOLER, which for the first time enables the theoretically guaranteed global optimization of complex non-convex problems; Liang et al. [32] proposed a multi-objective evolutionary NAS frame-work based on a weight-sharing supernet to improve the search efficiency of traditional evolutionary-computation-based NAS; Zhang et al. [33] progressed a remote sensing scene classification with a parameter learning algorithm-based co-evolutionary. Esteban et al. [34] proposed a parallel evolutionary algorithm for image classification, which optimized both network architectures and hyperparameters, demonstrating significant improvements in classification accuracy. Furthermore, in 2023, Abdullah et al. [35] introduced an Evolutionary Deep Convolutional Neural Network System (EDCNNS) for Alzheimer’s disease (AD) diagnostic systems, achieving substantial reduction in computational latency while enhancing diagnostic accuracy.

3. Proposed Methods

3.1. Overall Framework

The overall framework of the proposed water body identification is illustrated in Figure 1. It mainly consists of three main modules: (1) Hyperparameter optimization: evolutionary algorithms is used to optimize model hyperparameters, generating the optimal settings through a hybrid strategy. (2)Model training: An encoder–decoder architecture is used to train the water segmentation model. (3) Evaluation: A weighted loss function combining classification loss and Dice Loss is used for evaluation.
During image preprocessing, the model conducts data augmentation and normalization. The normalization scales the input image pixel values from [0, 255] to the range [0, 1] in order to accelerate model convergence and training speed. Data augmentation is performed via random scaling, flipping, and color transformation to enhance data diversity and improve the model’s generalization capability. Once preprocessing is complete, the processed image data is converted into PyTorch 2.3.0 tensors and transferred to the GPU. The bundled tensors are then fed into the U-Net model for training.
Prior to training the model, it is necessary to use an evolutionary algorithm to search for the best combination of hyperparameters. A search space containing the hyperparameters to be optimized, that is, a hyperparameter pool, is constructed manually. This pool includes key hyperparameters that significantly affect model performance, such as learning rate, batch size, and optimizer type. A combination of evolutionary algorithms is then selected. During the search process, the algorithm continuously generates new hyperparameter combinations according to a predetermined strategy. For each new hyperparameter combination, the evolutionary algorithm activates a fitness evaluation mechanism. This evaluation is performed by applying the hyperparameter combination to the U-Net model and conducting training and validation on a designated validation dataset. The model’s performance on the validation dataset, such as classification accuracy, segmentation precision, and convergence speed, is quantified as a fitness value. A higher fitness value indicates that the hyperparameter combination enables the model to perform better in the current task. Following the survival-of-the-fittest principle, the evolutionary algorithm selects hyperparameter combinations with excellent performance until a termination condition is met, finally outputting the optimal configuration.

3.2. Backbone

3.2.1. Network Model Selection

Common neural network models used for image identification include basic convolutional neural networks, deep training optimized networks, and attention enhancement networks [36]. A classical convolutional neural network (CNN) [37] suffers from the loss of spatial information due to pooling operations; the absence of a contextual information fusion mechanism renders it insufficiently sensitive to complex backgrounds such as the water and land boundary; and single-scale feature extraction proves inadequate for adapting to multi-scale objects [38]. The traditional deep training optimization network ResNet lacks channel attention and does not explicitly model inter-channel relationships [39], which may lead to the neglect of important spectral features; moreover, it is not suitable for pixel-level tasks because an effective upsampling mechanism is missing when it is directly employed for segmentation. The classical attention enhancement network SENet mainly focuses on channel weighting without explicitly optimizing spatial features [40], which potentially causes the loss of edge details; in addition, the absence of a multi-scale feature fusion mechanism makes it difficult to adapt to complex scenes.
In contrast to the aforementioned networks, the improved U-Net proposed in this work adopts a backbone structure that employs ResNet50 as the encoder. The use of residual structures alleviates gradient vanishing and facilitates the extraction of deeper and more discriminative features. This enhancement improves the capability to capture complex textures such as water ripples and turbid regions. Simultaneously the improved U-Net introduces an attention optimization mechanism by integrating the CBAM module and designing a dynamic gating fusion unit to achieve the synergistic optimization of spectral and spatial features. Furthermore, the improved U-Net employs a loss function optimization strategy. In response to the significant disparities in water body proportions typical of remote sensing imagery, a weighted multi-loss function is constructed. By jointly applying cross entropy loss, Dice Loss, and Focal Loss, the class imbalance are alleviated and the overall robustness of the model is improved.

3.2.2. The Network Structure

Traditional U-Net architectures generally employ simple convolution layers and pooling layers in the encoder, resulting in relatively simple structures. In the improved U-Net model, a more complex pre-trained network, ResNet50, is selected as the backbone. The output feature maps from ResNet50 are utilized in the decoder. By fusing these via skip connections with the decoder’s feature maps, high-resolution feature information is recovered, and water body features are accurately recognized. In the U-Net model, ResNet50 serves as the encoder responsible for extracting features from the input image. Specifically, feature maps extracted at different levels of ResNet50 are transmitted to the decoder via skip connections. This design not only preserves the strong feature extraction capability of ResNet50 but also achieves high-precision pixel-level segmentation through the U-Net decoder structure.
ResNet50 extracts multi-scale feature information from the input image using its residual modules. Each residual module is described by (1):
ResidualBlock ( x ) ReLU ( x + Con v 3 ( ReLU ( Con v 2 ( ReLU ( Con v 1 ( x ) ) ) ) ) )
where three convolution layers (Conv1 Conv2, and Conv3) are applied sequentially with the ReLU activation function, a skip connection, denoted as x, bypasses the convolutional layers and directly adds the input to the output, facilitating gradient flow and improving training efficiency.
In the decoder of U-Net, the feature maps extracted by ResNet50 are fused with the decoder feature maps via skip connections. This skip connection is implemented as shown in (2):
SkipConnection x , y = Concat UpSample x , y
where x represents the feature map extracted by ResNet50, y is the decoder feature map, Upsample indicates the upsampling operation, and Concat the concatenation operation. In the decoder, the fused feature maps are further processed by convolution layers and activation functions for feature fusion. This fusion is expressed in (3):
FeatureFusion x = ReLU Conv ReLU Conv x
where Conv signifies a convolution layer and ReLU denotes an activation function.

3.2.3. Dynamic Parameter Adjustment

We introduce the Dynamic Gate mechanism to achieve adaptive adjustment of attention weights through feature complexity sensing. This module is embedded in the enhanced CBAM (Enhanced CBAM) and integrated into the feature fusion layer (unetUp module) of the U-Net decoder. The overall architecture comprises three core components:
  • Spectral Attention Layer: Uses multi-scale one-dimensional convolution to capture cross-channel spectral dependency.
  • Spatial Attention Layer: Based on direction-aware convolution to extract spatial contextual feature.
  • Dynamic Gating Layer: Generates spatially adaptive attention intensity and feature retention coefficients.
In the dynamic adjustment mechanism, we assume the input feature is X R B × C × H × W . The dynamic parameter adjustment process follows the mathematical principles below.
First, the spectral weight calculation is performed. After compressing the spatial dimensions using global average pooling, cascaded 1D convolutional layers (with kernel sizes of 5 and 3, respectively) are applied to extract cross-channel spectral dependencies. The large convolution kernel captures global spectral trends, while the small kernel focuses on local band interactions. Its mathematical expression is given as (4). After this operation, the model automatically enhances the weights for water-sensitive bands such as near-infrared (NIR) and suppresses interference from visible light bands in shadow regions. The model’s noise robustness is further enhanced by smoothing out abnormal reflectance fluctuations induced by sensor noise
W s = s f 3 × 1 d f 5 × 1 GAP ( X )
For the second step, spatial weight generation, it is necessary to fuse average pooling (to retain the overall distribution) and max pooling (to highlight salient features). Orthogonally separated 7 × 1 and 1 × 7 convolution kernels are used to extract spatial context in the horizontal and vertical directions, respectively:
W p = σ f 7 v f 7 h AvgPool ( X ) ; MaxPool ( X )
At this stage, the horizontal convolution kernel reinforces the continuity of linear targets such as rivers and canals, while the vertical convolution kernel enhances the contrast between lake boundaries and land, thereby strengthening directional awareness. Pooling operations couple features from different receptive fields, enhancing the aggregated representation of fragmented water bodies such as pond clusters.
Finally, during the dynamic mixing output, spatially adaptive adjustment coefficients, including attention intensity ( γ ) and feature retention rate ( β ), are generated based on feature complexity. These coefficients are obtained through a lightweight network, as shown in (6):
[ γ , β ] = Γ complexity ( x )
where Γ complexity is constructed by global average pooling followed by two convolution layers with a compression ratio of 16:1, and the final output complies with a specific constraint. The feature fusion is then formulated as in (7):
X out = γ ( W p X ) + β X
After the dynamic mixed output operation, in regions with high complexity, the attention intensity γ increases and the feature retention rate β decreases, thereby enhancing the characterization of boundary features by the attention mechanism γ . In contrast, in relatively simple regions, the attention intensity β decreases and the feature retention rate increases, which helps preserve low-level spectral features and prevents over-smoothing. The design of the residual connection ensures that gradients can directly bypass to the original features, mitigating gradient attenuation problems in deep networks. The process of the dynamic parameter adjustment mechanism described above is illustrated in Figure 2.

3.3. Hyperparameter Tuning Combined with Multiple Evolutionary Algorithms

To overcome the inherent trade-off between global exploration and local convergence in single-method optimization, this study introduces an integrated hyperparameter optimization framework combining Genetic Algorithm (GA), Ant Colony Optimization (ACO), Particle Swarm Optimization (PSO), and the Alpha Evolutionary (AE) algorithm. Each algorithm is assigned a distinct role according to its operational principles: GA performs wide-ranging exploration through selection, crossover, and mutation to preserve population diversity; ACO employs a pheromone-mediated positive feedback mechanism to intensify search trajectories in promising regions; PSO accelerates convergence by steering particles based on individual and collective experience; and AE concentrates on refining the elite solution space during later stages to secure stable, high-precision outcomes. Orchestrating their collaboration, a dynamic weight fusion mechanism adaptively modulates the contribution of each algorithm in response to iterative progress and performance feedback, thereby harmonizing global exploration, path refinement, rapid convergence, and solution-space polishing to achieve fully automated and near-optimal hyperparameter configuration.

3.3.1. Hyperparameter Optimization

The hyperparameter optimization in this work refers to the optimization of the learning rate, batch size, and momentum for the improved U-Net model. After initializing the hyperparameter values using an evolutionary algorithm, the hyperparameter array is input into the U-Net model for prediction. The search terminates after a specified number of iterations. All evaluation results of water body identification are then ranked from best to worst, and the hyperparameter array with the optimal evaluation result is selected for subsequent formal water body identification.
In this study, the evaluation function used in the hyperparameter optimization module is the F1_Score.
F 1 _ Score = 2 Precision Recall Precision + Recall
Precision = TP TP + FP , Recall = TP TP + FN
( x LearningRate , x BatchSize , x Momentum ) * = argmax   f F 1 _ Score x LearningRate , x BatchSize , x Momentum x LearningRate X LearningRate , x BatchSize X BatchSize , x Momentum X Momentum
Its mathematical expression is given in (8), and the associated parameter expressions are provided in (9). Therefore, the hyperparameter optimization problem is defined by the mathematical model expressed in (10), where ( x LearningRate , x BatchSize , x Momentum ) * represents the hyperparameter configuration that produces the optimal value f F 1 _ Score . The hyperparameter set x LearningRate , x BatchSize , x Momentum can take any value from the search space X LearningRate , X BatchSize , X Momentum .

3.3.2. Weight Adaptive Optimization Based on Multi-Algorithm Fusion

In view of the tendency of a single evolutionary algorithm to fall into local optima and its limited convergence speed, this work proposes a four-stage joint optimization strategy that integrates Genetic Algorithm (GA), Ant Colony Optimization (ACO), Particle Swarm Optimization (PSO), and Alpha Evolutionary Algorithm (AE) [41]. Through the construction of a dynamic weight fusion mechanism and a layered optimization framework, this strategy combines the dual advantages of global search and local refinement to achieve collaborative exploration of the hyperparameter space as well as fine adjustment in the solution space.
The core innovations of this strategy are reflected in the following four aspects:
Multi-Algorithm Joint Prediction Model: Assume that the outputs of the U-Net model optimized by GA, ACO, PSO, and AE are P GA , P ACO , P PSO , and P AE , respectively. The joint prediction result is generated by weighted fusion as expressed in (11):
P Ensemble = ω GA × P GA + ω ACO × P ACO + ω PSO × P PSO + ω AE × P AE
where ω i = 1 , the weight coefficients are dynamically adjusted. This model integrates the population diversity of GA, the positive feedback path optimization of ACO, the gradient-driven characteristics of PSO, and the advanced comprehensive optimization capability of AE, thereby forming a balanced architecture of global exploration and local refinement.
The dynamic weight update mechanism significantly impacts the final performance of the model, making the fine-tuning of joint weights a critical step. By dynamically adjusting the weights to achieve optimal results, the process can be divided into three stages. In the early iteration stage (T1), the weight of the GA (Genetic Algorithm) is increased to approximately 0.35 ± 0.05, which enhances the global topological search capability and prevents premature convergence. In the middle iteration stage (T2), a balance is dynamically achieved between the ACO (Ant Colony Optimization) weight, set to approximately 0.28, and the PSO (Particle Swarm Optimization) weight, set to approximately 0.32. This forms a dual driving mechanism based on path optimization (via ACO) and gradient-driven refinement (via PSO), enabling efficient exploration and exploitation of the solution space. Finally, in the late-convergence stage (T3), the AE (Autoencoder) weight is increased to approximately 0.4 ± 0.06, activating the elite solution space compression mechanism to refine and converge on the optimal result. Table 1 descripts their individual functions.

3.4. Loss Function

To jointly optimize classification confidence and geometric consistency, this work proposes a linear weighted loss fusion strategy. It combines the classification loss (either Cross-Entropy Loss or Focal Loss) with Dice Loss to form a composite loss function as shown in (12):
L Total = λ × L cls + β × L dice
where L cls denotes the classification loss (either L CE or L Focal ), L Dice represents Dice Loss, and λ and β are dynamic weighting coefficients. The default settings are λ = 1 and β = 1 . These coefficients can be adaptively adjusted during training using evolutionary algorithms (such as Genetic Algorithm or Particle Swarm Optimization) to balance the contributions of the different loss components in model optimization. This strategy, through end-to-end joint training, enables simultaneous optimization of the following objectives during backpropagation:
Optimization of Classification Confidence: Constraining the model output probability distribution to match the true labels to enhance pixel-level classification accuracy.
Optimization of Geometric Consistency: Improving the topological alignment between the predicted regions and the ground truth mask to refine boundary segmentation performance.

4. Experiments and Results

4.1. Datasets

The datasets were obtained from two sources. One dataset was published on Kaggle.com by Francisco Escobar for training water body identification models. In this dataset each image is provided with a black-and-white mask in which white represents water and black represents all non-water areas. These masks are generated through the computation of NDWI (Normalized Water Difference Index). NDWI is frequently used to detect and measure vegetation in satellite imagery but a higher threshold is employed for water detection.
The second dataset consists of atmospherically corrected multispectral Sentinel-2 images [40]. The Sentinel-2 dataset provides continuous coverage for global land surface monitoring and offers several advantages compared with other image types such as high spatial resolution and free data access. Because the input to the model must conform to the VOC format, the masks are required to set all non-target objects to a pixel value of 0 and the target objects to a pixel value of 1. To meet this requirement, the masks in this dataset are converted to the corresponding format through a script, thereby ensuring that the model correctly identifies water body areas and performs water body identification with higher accuracy.
During preprocessing, we used the Kaggle dataset as supplemental training data for the Sentinel-2 dataset to prevent model overfitting and improve generalization and stability. The primary dataset, which served as the basis for model training and evaluation, was manually digitized to generate ground truth. For areas where visual assessment was challenging, we referred to auxiliary high-resolution imagery from sources like Google Earth for accurate fine-scale identification. The two datasets were merged during the pre-processing phase after being split into 512 × 512 patches, which served as the model input. The combined dataset was then partitioned into 70% for training, 15% for validation, and 15% for testing.

4.2. Evaluation Metrics

This study employs five evaluation metrics to assess the experimental results. These metrics include Pixel Accuracy (PA), F1_Score, Mean Intersection-over-Union (mIoU), Mean Pixel Accuracy (mPA), and Mean Precision (mPrecision).
Pixel Accuracy (PA) reflects the overall capability of the model to distinguish water pixels from non-water pixels. Its value ranges from 0 to 1; a value closer to 1 indicates a stronger ability to correctly classify true water bodies. The mathematical expression for PA is given in (13).
PA = TP + TN TP + TN + FP + FN
F1_Score evaluates the comprehensive performance of the model on imbalanced datasets. Its value ranges from 0 to 1. F1 is the harmonic mean of precision and recall; a value closer to 1 signifies superior performance on imbalanced data. The mathematical expression for the F1_Score provided in above (8).
Mean Intersection-over-Union (mIoU) quantifies the geometric consistency between the predicted regions and the true regions. The value ranges from 0 to 1; a higher mIoU signifies better overlap between prediction and truth and indicates more precise boundary segmentation. The mathematical expression for mIoU is given in (14):
mIoU = 1 C c = 1 C T P c T P c + F P c + F N c
where C represents the class and two classes are used (water and background). Mean Pixel Accuracy (mPA) measures the classification accuracy of each class independently. Its value ranges from 0 to 1; a higher mPA indicates that the model exhibits stronger robustness against local noise and provides better suppression of false detections in localized areas. The mathematical expression for mPA is shown in (15).
mPA = 1 C c = 1 C T P c T P c + F N c
Mean Precision (mPrecision) assesses the reliability of positive predictions. Its value ranges from 0 to 1; a higher mPrecision indicates that a larger proportion of samples predicted as positive are indeed positive, implying a lower false alarm rate. The mathematical expression for mPrecision is provided in (16).
mPrecision = 1 C c = 1 C T P c T P c + F P c
Collectively these evaluation metrics assess the performance of the water body identification model from multiple perspectives. PA and F1_Score validate overall global performance, whereas mIoU and mPA provide a dual evaluation of global geometric accuracy and local classification accuracy, and mPrecision primarily constrains the false positive rate.

4.3. Results and Analysis

Baseline models used for comparison include U-Net variants that incorporate different backbone networks. In this work the backbone networks comprise VGG16, DenseNet, and MobileNet. Two optimizers, namely the SGD optimizer and the Adam optimizer, are also compared.
To objectively compare the prediction results of the improved U-Net model with those of the baseline model the hyperparameters for the baseline models are set to be identical to those of the experimental model. These parameters include the learning rate, optimizer type, batch size, and the number of training epochs. Pre-trained weights are used to initialize the backbone network for each model. All network models paired with each backbone network are integrated within the U-Net architecture. All eight backbone-optimizer combinations, ResNet50 + Adam, ResNet50 + SGD, VGG16 + Adam, VGG16 + SGD, DenseNet + Adam, DenseNet + SGD, MobileNet + Adam, and MobileNet + SGD, used identical hyperparameters: a maximum learning rate of 1 × 10−4, minimum learning rate of 1 × 10−6, and 200 training batches.
To investigate the superiority of the proposed model, eight combinations formed by four different backbone networks and two optimizers are compared using the five-evaluation metrics mentioned above. The evaluation results are presented in Figure 3 and Table 2.
It is observed that the combination of the ResNet backbone with the Adam optimizer exhibits a more stable training process compared with the other seven baseline models. During the iterative process both PA and F1_Score remain higher than those obtained from the other seven models. After the final iteration the best PA and best F1_Score are significantly higher than those from the other baseline models. With the exception of the ResNet combined with the SGD optimizer the ResNet + Adam combination outperforms the other models by more than 10 percent. This demonstrates the superiority of the ResNet + Adam combination, and the visualization results can be seen in Figure 4, where we can intuitively observe that the water body extraction result obtained by the combination of ResNet + Adam is closest to the original image and the ground truth, especially in water body edge details and small water bodies.

4.4. Ablation Experiment

Ablation experiments demonstrate the effects of evolutionary algorithms, attention modules, linear weighted loss function fusion, and evolutionary algorithms on water extraction from remote sensing images. It can be observed that after incorporating the CBAM attention mechanism the model attains a PA value exceeding 0.965, which is well above the acceptable threshold of 0.90. The F1_Score exceeds 0.945, surpassing the minimal acceptable value of 0.85. Similarly, mIoU reaches above 0.902, mPA exceeds 94.4, and mPrecision reaches above 0.952, all of which are within excellent ranges. Compared with the U-Net model without the CBAM attention mechanism, the network with CBAM consistently outperforms across all evaluation metrics and exhibits a more stable training process. This demonstrates the effectiveness of the proposed strategy. To assess the effectiveness of loss function fusion for enhancing water body identification performance the model employing the unfused cross entropy loss and Focal loss is compared with the model using the fused Focal + Dice loss function. The comparison results are visualized, and evaluation results are provided in Table 3.
After fusing the loss functions the model attains a PA exceeding 0.965, an F1_Score above 0.945, an mIoU above 0.902, an mPA exceeding 0.944, and an mPrecision above 0.952. Apart from the mPrecision metric the network employing the fused loss functions outperforms the networks using only cross entropy loss or Focal loss. This further verifies the effectiveness of the proposed strategy.
To examine the effectiveness of the joint evolutionary algorithm for hyperparameter optimization, the evaluation metrics obtained from four individual evolutionary algorithms are compared with those achieved using the joint evolutionary algorithm. The results are shown in Table 4 and visualized as shown in Figure 5.
It indicates that the model optimized by the joint algorithm outperforms the models optimized by any single evolutionary algorithm in all evaluation metrics. Furthermore, the evaluation results demonstrate that the Alpha Evolutionary Algorithm yields superior performance compared with the other three evolutionary algorithms. This indirectly confirms that assigning a higher weight to the Alpha Evolutionary Algorithm in the joint weighting strategy is appropriate. The water body comparison results are shown in Figure 6.

5. Conclusions

This study addressed the challenges of automating water body identification from satellite imagery by introducing a U-Net model optimized through a hybrid evolutionary algorithm. The core of our approach lies in the integration of three key components: (1) an enhanced U-Net with a ResNet50 encoder and a dynamic attention mechanism for improved multi-scale feature extraction and adaptive feature calibration; (2) a hybrid evolutionary strategy combining the Genetic Algorithm, Ant Colony Optimization, Particle Swarm Optimization, and the Alpha Evolutionary algorithm to automate hyperparameter configuration, mitigating the reliance on manual tuning; and (3) a learnable multi-loss fusion strategy to balance classification and geometric consistency, effectively addressing class imbalance. Extensive experiments on the Kaggle and Sentinel-2 datasets demonstrated the efficacy of the proposed method. Our model achieved a Pixel Accuracy of 96.79% and an F1-Score of 0.9475, outperforming various baseline models by a considerable margin (over 10% in key metrics). Ablation studies confirmed the contribution of each component, showing that the hybrid evolutionary optimizer yielded better performance than any single algorithm alone.
Despite these results, we note two primary limitations. First, the model’s training relied on publicly available Sentinel-2 imagery, and its performance on data from other sensors or under significantly different imaging conditions requires further validation. Second, the joint optimization of multiple evolutionary algorithms is computationally intensive, leading to increased training time and resource demands. Future work will focus on several directions to enhance the framework’s practicality and scope. This includes incorporating multi-source remote sensing data (e.g., Landsat, GF series, SAR) to improve generalization, exploring lightweight evolutionary strategies or surrogate models to reduce computational costs, and extending the model’s capability to multi-class land-cover segmentation and temporal analysis for dynamic monitoring. The research framework presented here offers a viable path toward fully automated remote sensing image analysis, with potential applications in water resource management, environmental monitoring, and disaster assessment.

Author Contributions

Conceptualization, Y.Y. and P.W.; methodology, Y.Y. and Z.Q.; software, Z.Q. and Z.L.; validation, Y.Y. and Z.Q.; formal analysis, Y.Y.; investigation, J.G.; resources, X.D.; data curation, T.C. and J.Z.; writing—original draft preparation, Y.Y.; writing—review and editing, Y.Y., Z.Q. and T.C.; visualization, Z.L. and X.D.; supervision, Y.Y., P.W. and Z.L.; project administration, P.W. and Z.Q.; funding acquisition, Y.Y., Z.L. and J.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Funded Postdoctoral Research Program (GZC20241900), Chongqing Natural Science Foundation (CSTB2024JXJL-YFX0079), the open project of Dazhou Key Laboratory of Government Data Security under Grant (ZSAQ202501, ZSAQ202502, ZSAQ202507, ZSAQ202505), Key Laboratory of Remote Sensing Application and Innovation (LRSAI-2025004), 2024 Chongqing Municipality Special Funding for Postdoctoral Research Projects(2024CQBSHTB3045), Sichuan University students innovation and entrepreneurship training program(S202510621042).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The study did not report any data.

Acknowledgments

Great thanks to the editors and all the anonymous reviewers for their constructive and insightful comments on the earlier drafts of this manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Li, A.; Jiao, L.; Zhu, H.; Li, L.; Liu, F. Multitask Semantic Boundary Awareness Network for Remote Sensing Image Segmentation. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5400314. [Google Scholar] [CrossRef]
  2. Dong, Z.; Wang, G.; Amankwah, S.O.Y.; Wei, X.; Hu, Y.; Feng, A. Monitoring the Summer Flooding in the Poyang Lake Area of China in 2020 Based on Sentinel-1 Data and Multiple Convolutional Neural Networks. Int. J. Appl. Earth Obs. Geoinf. 2021, 102, 102400. [Google Scholar] [CrossRef]
  3. Wambugu, N.; Chen, Y.; Xiao, Z.; Wei, M.; Bello, S.A.; Junior, J.M.; Li, J. A Hybrid Deep Convolutional Neural Network for Accurate Land Cover Classification. Int. J. Appl. Earth Obs. Geoinf. 2021, 103, 102515. [Google Scholar] [CrossRef]
  4. He, C.; Sadeghpour, H.; Shi, Y.; Mishra, B.; Roshankhah, S. Mapping Distribution of Fractures and Minerals in Rock Samples Using Res-VGG-UNet and Threshold Segmentation Methods. Comput. Geotech. 2024, 175, 106675. [Google Scholar] [CrossRef]
  5. McFeeters, S.K. The Use of the Normalized Difference Water Index (NDWI) in the Delineation of Open Water Features. Int. J. Remote Sens. 1996, 17, 1425–1432. [Google Scholar] [CrossRef]
  6. Xu, H. Modification of Normalised Difference Water Index (NDWI) to Enhance Open Water Features in Remotely Sensed Imagery. Int. J. Remote Sens. 2006, 27, 3025–3033. [Google Scholar] [CrossRef]
  7. Feyisa, G.L.; Meilby, H.; Fensholt, R.; Proud, S.R. Automated Water Extraction Index: A New Technique for Surface Water Mapping Using Landsat Imagery. Remote Sens. Environ. 2014, 140, 23–35. [Google Scholar] [CrossRef]
  8. Maksimovic, V.; Jaksic, B.; Milosevic, M.; Todorovic, J.; Mosurovic, L. Comparative Analysis of Edge Detection Operators Using a Threshold Estimation Approach on Medical Noisy Images with Different Complexities. Sensors 2024, 25, 87. [Google Scholar] [CrossRef] [PubMed]
  9. Kılıç, Z.A.; Kaya, K.; Yener, A.K. A Data Mining-Based Daylighting Design Decision Support Model for Achieving Visual Comfort Conditions in the Multi-Functional Residential Space. J. Build. Eng. 2025, 103, 112141. [Google Scholar] [CrossRef]
  10. Zhang, Y.; Ming, D.; Dong, D.; Xu, L. Object-Oriented U-GCN for Open-Pit Mining Extraction from High Spatial Resolution Remote-Sensing Images of Complex Scenes. Int. J. Remote Sens. 2024, 45, 8313–8333. [Google Scholar] [CrossRef]
  11. Isaksson, J.; Arvidsson, I.; Åaström, K.; Heyden, A. Semantic Segmentation of Microscopic Images of H&E Stained Prostatic Tissue Using CNN. In Proceedings of the International Joint Conference on Neural Networks, Anchorage, AK, USA, 14–19 May 2017; pp. 1252–1256. [Google Scholar]
  12. Long, J.; Shelhamer, E.; Darrell, T. Fully Convolutional Networks for Semantic Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  13. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  14. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  15. Howard, A.; Sandler, M.; Chu, G.; Chen, L.-C.; Chen, B.; Tan, M.; Wang, W.; Zhu, Y.; Pang, R.; Vasudevan, V.; et al. Searching for MobileNetV3. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 1314–1324. [Google Scholar]
  16. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  17. Wang, G.; Wu, M.; Wei, X.; Song, H. Water Identification from High-Resolution Remote Sensing Images Based on Multidimensional Densely Connected Convolutional Neural Networks. Remote Sens. 2020, 12, 794. [Google Scholar] [CrossRef]
  18. Wagner, F.; Eltner, A.; Maas, H.-G. River Water Segmentation in Surveillance Camera Images: A Comparative Study of Offline and Online Augmentation Using 32 CNNs. Int. J. Appl. Earth Obs. Geoinf. 2023, 119, 103305. [Google Scholar] [CrossRef]
  19. Rasley, J.; Rajbhandari, S.; Ruwase, O.; He, Y. DeepSpeed: System Optimizations Enable Training Deep Learning Models with Over 100 Billion Parameters. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Virtual Event, CA, USA, 6–10 July 2020; pp. 3505–3506. [Google Scholar]
  20. Kingma, D.P.; Ba, J.L. Adam: A Method for Stochastic Optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  21. Zhang, L.; Li, Y.; Hou, Z.; Li, X.; Geng, H.; Wang, Y.; Li, J.; Zhu, P.; Mei, J.; Jiang, Y.; et al. Deep Learning and Remote Sensing Data Analysis. Geomat. Inf. Sci. Wuhan Univ. 2020, 45, 1857–1864. [Google Scholar] [CrossRef]
  22. Sun, D.; Gao, G.; Huang, L.; Liu, Y.; Liu, D. Extraction of water bodies from high-resolution remote sensing imagery based on a deep semantic segmentation network. Sci. Rep. 2024, 14, 14604. [Google Scholar] [CrossRef]
  23. Feng, S.; Yang, Q.; Jia, W.; Wang, M.; Liu, L. Information Extraction of Inland Surface Water Bodies Based on Optical Remote Sensing: A Review. Remote Sens. Nat. Resour. 2024, 36, 41–56. [Google Scholar] [CrossRef]
  24. Wen, Q.; Li, L.; Xiong, L.; Du, L.; Liu, Q.; Wen, Q. A Review of Water Body Extraction from Remote Sensing Images Based on Deep Learning. Remote Sens. Nat. Resour. 2024, 36, 57–71. [Google Scholar] [CrossRef]
  25. Tomczyk, M.K.; Kadziński, M. Decomposition-Based Interactive Evolutionary Algorithm for Multiple Objective Optimization. IEEE Trans. Evol. Comput. 2020, 24, 320–334. [Google Scholar] [CrossRef]
  26. Ye, T.; Zhang, X.; Wang, C.; Jin, Y. An Evolutionary Algorithm for Large-Scale Sparse Multiobjective Optimization Problems. IEEE Trans. Evol. Comput. 2020, 24, 380–393. [Google Scholar] [CrossRef]
  27. Holland, J.H. Adaptation in Natural and Artificial Systems; MIT Press: Cambridge, MA, USA, 1992. [Google Scholar]
  28. Dorigo, M.; Gambardella, L.M. Ant Colony System: A Cooperative Learning Approach to the Traveling Salesman Problem. IEEE Trans. Evol. Comput. 1997, 1, 53–66. [Google Scholar] [CrossRef]
  29. Dorigo, M.; Maniezzo, V.; Colorni, A. Ant System: Optimization by a Colony of Cooperating Agents. IEEE Trans. Syst. Man Cybern. Part B Cybern. 1996, 26, 29–41. [Google Scholar] [CrossRef]
  30. Kennedy, J.; Eberhart, R. Particle Swarm Optimization. In Proceedings of the IEEE International Conference on Neural Networks, Perth, Australia, 27 November–1 December 1995; pp. 1942–1948. [Google Scholar]
  31. Li, B.; Wei, Z.; Wu, J.; Yu, S.; Zhang, T.; Zhu, C.; Zheng, D.; Guo, W.; Zhao, C.; Zhang, J. Machine Learning-Enabled Globally Guaranteed Evolutionary Computation. Nat. Mach. Intell. 2023, 5, 123–134. [Google Scholar] [CrossRef]
  32. Liang, J.; Zhu, K.; Li, Y.; Li, Y.; Gong, Y. Multi-Objective Evolutionary Neural Architecture Search with Weight-Sharing Supernet. Appl. Sci. 2024, 14, 6143. [Google Scholar] [CrossRef]
  33. Zhang, D.; Zhou, Y.; Zhao, J.; Zhou, Y. Co-Evolution-Based Parameter Learning for Remote Sensing Scene Classification. Remote Sens. 2021, 13, 4125. [Google Scholar] [CrossRef]
  34. Real, E.; Moore, S.; Selle, A.; Saxena, S.; Suematsu, Y.L.; Tan, J.; Le, Q.V.; Kurakin, A. Large-Scale Evolution of Image Classifiers. In Proceedings of the 34th International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; pp. 2902–2911. [Google Scholar]
  35. Lakhan, A.; Grønli, T.M.; Muhammad, G.; Tiwari, P. EDCNNS: Federated Learning Enabled Evolutionary Deep Convolutional Neural Network for Alzheimer Disease Detection. Appl. Soft Comput. 2023, 147, 110804. [Google Scholar] [CrossRef]
  36. Xiang, D.; Zhang, X.; Wu, W.; Liu, H. DensePPMUNet-a: A robust deep learning network for segmenting water bodies from aerial images. IEEE Trans. Geosci. Remote Sens. 2023, 61, 4202611. [Google Scholar] [CrossRef]
  37. Liu, W.; Chen, X.; Ran, J.; Liu, L.; Wang, Q.; Xin, L.; Li, G. LaeNet: A Novel Lightweight Multitask CNN for Automatically Extracting Lake Area and Shoreline from Remote Sensing Images. Remote Sens. 2021, 13, 56. [Google Scholar] [CrossRef]
  38. Luo, X.; Hu, Z.; Liu, L. Investigating the seasonal dynamics of surface water over the Qinghai–Tibet Plateau using Sentinel-1 imagery and a novel gated multiscale ConvNet. Int. J. Digit. Earth 2023, 16, 1372–1394. [Google Scholar] [CrossRef]
  39. Li, M.; Wu, P.; Wang, B.; Park, H.; Yang, H.; Wu, Y. A Deep Learning Method of Water Body Extraction From High Resolution Remote Sensing Images With Multisensors. IEEE J-STARS 2021, 14, 3120–3132. [Google Scholar] [CrossRef]
  40. Dai, X.; Xia, M.; Weng, L.; Hu, K.; Lin, H.; Qian, M. Multiscale Location Attention Network for Building and Water Segmentation of Remote Sensing Image. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5609519. [Google Scholar] [CrossRef]
  41. Gao, H.; Zhang, Q. Alpha evolution: An efficient evolutionary algorithm with evolution path adaptation and matrix generation. Eng. Appl. Artif. Intell. 2024, 137, 109202. [Google Scholar] [CrossRef]
Figure 1. The water body identification framework based on hybrid evolutionary algorithms.
Figure 1. The water body identification framework based on hybrid evolutionary algorithms.
Biomimetics 10 00732 g001
Figure 2. Dynamic Parameter Adjustment.
Figure 2. Dynamic Parameter Adjustment.
Biomimetics 10 00732 g002
Figure 3. Comparison between the U-Net model with the Adam optimizer and other baseline models.
Figure 3. Comparison between the U-Net model with the Adam optimizer and other baseline models.
Biomimetics 10 00732 g003
Figure 4. Comparison of water extraction results with different backbone network and optimizer combinations.
Figure 4. Comparison of water extraction results with different backbone network and optimizer combinations.
Biomimetics 10 00732 g004
Figure 5. Comparison of Evaluation Metrics for Evolutionary Algorithms and the Combined Algorithm.
Figure 5. Comparison of Evaluation Metrics for Evolutionary Algorithms and the Combined Algorithm.
Biomimetics 10 00732 g005
Figure 6. Comparison of water body segmentation results with different evolutionary algorithm.
Figure 6. Comparison of water body segmentation results with different evolutionary algorithm.
Biomimetics 10 00732 g006
Table 1. Parameters and tasks involved in each evolutionary algorithm.
Table 1. Parameters and tasks involved in each evolutionary algorithm.
AlgorithmParameter and TaskInitial Value
GAPopulation Size: Determines the diversity of potential hyperparameter sets in the gene pool. 20
Crossover Rate: Controls the probability of combining genetic material from two parents to produce offspring.0.7
Mutation Rate: Introduces random changes to offspring, maintaining population diversity and preventing premature convergence to local optima.0.1
Selection Method: Governs which individuals are chosen for reproduction. Tournament selection provides good selective pressure.Tournament (size = 3)
ACONumber of Ants: Analogous to population size. Each ant constructs a solution (a path representing a hyperparameter set).20
Pheromone Influence (α): Determines the weight of pheromone trails in path selection. Higher values increase exploitation of known good paths.1.0
Heuristic Influence (β): Determines the weight of heuristic information in path selection. Higher values favor exploration.2.0
Pheromone Evaporation Rate (ρ): Simulates the evaporation of pheromones over time, preventing unlimited accumulation and allowing the colony to forget poorer paths, facilitating exploration.0.5
PSONumber of Particles Similarly to population size. Each particle represents a potential hyperparameter set flying through the search space.20
Inertia Weight (w): Balances exploration (high w) and exploitation (low w). It controls the particle’s momentum based on its previous velocity.0.5
Cognitive Coefficient (c1): Attracts the particle towards its personal best position (pbest), encouraging exploitation of personally found good solutions.1.5
Social Coefficient (c2): Attracts the particle towards the swarm’s global best position (gbest), encouraging convergence towards the best-known solution.2.0
AEPopulation Size: Determines the number of candidate solutions maintained per iteration.20
Decay Coefficient of Disturbance: Controls the decaying speed of the disturbance intensity (alpha) during the algorithm’s iterative process0.9
Evaporation Rate: Simulates the pheromone evaporation mechanism like Ant Colony Optimization0.2
Table 2. Maximum Pixel Accuracy and F1_score For Each Backbone and Optimizer Combination.
Table 2. Maximum Pixel Accuracy and F1_score For Each Backbone and Optimizer Combination.
Backbone + OptimizerBest PA(%)Best F1_Score(%)
ResNet50 + Adam96.789794.75
ResNet50 + SGD94.472991.59
VGG16 + Adam81.244569.24
VGG16 + SGD78.610065.83
DenseNet + Adam82.877072.59
DenseNet + SGD81.865171.67
MobileNet + Adam83.362873.45
MobileNet + SGD81.936371.78
Table 3. Analysis of the Ablation Effects of Individual Components.
Table 3. Analysis of the Ablation Effects of Individual Components.
MethodsPA(%)F1_Score(%)mIoU(%)mPA(%)mPrecision(%)
U-Net96.2893.9088.8793.8494.66
U-Net +CELoss96.6794.5089.6793.8595.49
U-Net +FocalLoss96.7094.5089.9193.9095.46
U-Net +FocalLoss + Dice Loss + CBAM96.9394.9090.6194.6295.57
Table 4. Maximum Pixel Accuracy and F1_score For evolutionary algorithms.
Table 4. Maximum Pixel Accuracy and F1_score For evolutionary algorithms.
AlgorithmBest PA(%)Best F1Score(%)
Genetic Algorithm (GA)95.5892.21
Ant Colony Optimization (ACO)95.7993.15
Particle Swarm Optimization (PSO)96.1194.31
Alpha Evolutionary (AE)96.5395.57
The Joint: GA + ACO + PSO + AE97.6996.32
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yuan, Y.; Wei, P.; Qi, Z.; Deng, X.; Zhang, J.; Gan, J.; Chen, T.; Li, Z. Water Body Identification from Satellite Images Using a Hybrid Evolutionary Algorithm-Optimized U-Net Framework. Biomimetics 2025, 10, 732. https://doi.org/10.3390/biomimetics10110732

AMA Style

Yuan Y, Wei P, Qi Z, Deng X, Zhang J, Gan J, Chen T, Li Z. Water Body Identification from Satellite Images Using a Hybrid Evolutionary Algorithm-Optimized U-Net Framework. Biomimetics. 2025; 10(11):732. https://doi.org/10.3390/biomimetics10110732

Chicago/Turabian Style

Yuan, Yue, Peiyang Wei, Zhixiang Qi, Xun Deng, Ji Zhang, Jianhong Gan, Tinghui Chen, and Zhibin Li. 2025. "Water Body Identification from Satellite Images Using a Hybrid Evolutionary Algorithm-Optimized U-Net Framework" Biomimetics 10, no. 11: 732. https://doi.org/10.3390/biomimetics10110732

APA Style

Yuan, Y., Wei, P., Qi, Z., Deng, X., Zhang, J., Gan, J., Chen, T., & Li, Z. (2025). Water Body Identification from Satellite Images Using a Hybrid Evolutionary Algorithm-Optimized U-Net Framework. Biomimetics, 10(11), 732. https://doi.org/10.3390/biomimetics10110732

Article Metrics

Back to TopTop