Next Article in Journal
An Adaptive Sequential Phase Optimization Method Based on Coherence Stability Detection and Adjustment Correction
Previous Article in Journal
Evaluating Terrestrial Water Storage, Fluxes, and Drivers in the Pearl River Basin from Downscaled GRACE/GFO and Hydrometeorological Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Learning-Based Classification of Aquatic Vegetation Using GF-1/6 WFV and HJ-2 CCD Satellite Data

1
Key Laboratory of Digital Earth Science, Aerospace Information Research Institute, Chinese Academy of Sciences, Beijing 100094, China
2
International Research Center of Big Data for Sustainable Development Goals, Beijing 100094, China
3
University of Chinese Academy of Sciences, Beijing 100049, China
4
Satellite Application Center for Ecology and Environment, MEE, Beijing 100094, China
5
Jiangsu Tianyan Environment Technology Co., Ltd., Changzhou 213022, China
6
College of Resources Environment and Tourism, Capital Normal University, Beijing 100048, China
7
MOE Key Laboratory of 3D Information Acquisition and Application, Capital Normal University, Beijing 100048, China
8
Beijing Laboratory of Water Resource Security, Capital Normal University, Beijing 100048, China
9
State Key Laboratory Incubation Base of Urban Environmental Processes and Digital Simulation, Capital Normal University, Beijing 100048, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(23), 3817; https://doi.org/10.3390/rs17233817
Submission received: 17 October 2025 / Revised: 16 November 2025 / Accepted: 22 November 2025 / Published: 25 November 2025

Highlights

What are the main findings?
  • This study demonstrated that high-temporal-resolution Chinese satellite data—aofen-1/6 (GF-1/6) Wide Field of View (WFV) and Huanjing-2A/B (HJ-2A/B)—with 16 m spatial resolution, 1-day revisit capability after constellation networking, and consistent spectral performance, can support effective extraction of aquatic vegetation.
  • This study showed that deep learning algorithms can efficiently learn spectral–spatial information from multisource imagery, enabling accurate aquatic vegetation classification, and identified the optimal input parameters and network components for achieving the best performance.
What is the implication of the main finding?
  • This study provides a practical and reliable technical framework for monitoring aquatic vegetation during the peak growing season in the Yangtze River Basin, particularly under limited clear-sky conditions.
  • The study’s examination of optimal deep learning configurations for aquatic vegetation mapping offers a useful reference for future work in related fields.

Abstract

The Yangtze River Basin, one of China’s most vital watersheds, sustains both ecological balance and human livelihoods through its extensive lake systems. However, since the 1980s, these lakes have experienced significant ecological degradation, particularly in terms of aquatic vegetation decline. To acquire reliable aquatic vegetation data during the peak growing season (July–September), when clear-sky conditions are scarce, we employed Chinese domestic satellite imagery—Gaofen-1/6 (GF-1/6) Wide Field of View (WFV) and Huanjing-2A/B (HJ-2A/B) Charge-Coupled Device (CCD)—with approximately one-day revisit frequency after constellation networking, 16 m spatial resolution, and excellent spectral consistency, in combination with deep learning algorithms, to monitor aquatic vegetation across the basin. Comparative experiments identified the near-infrared, red, and green bands as the most informative input features, with an optimal input size of 256 × 256. Through visual interpretation and dataset augmentation, we generated a total of 5016 labeled image pairs of this size. The U-Net++ model, equipped with an EfficientNet-B5 backbone, achieved robust performance with an mIoU of 90.16% and an mPA of 95.27% on the validation dataset. On independent test data, the model reached an mIoU of 79.10% and an mPA of 86.42%. Field-based assessment yielded an overall accuracy (OA) of 75.25%, confirming the reliability of the model. As a case study, the proposed model was applied to satellite imagery of Lake Taihu captured during the peak growing season of aquatic vegetation (July–September) from 2020 to 2025. Overall, this study introduces an automated classification approach for aquatic vegetation using 16 m resolution Chinese domestic satellite imagery and deep learning, providing a reliable framework for large-scale monitoring of aquatic vegetation across lakes in the Yangtze River Basin during their peak growth period.

1. Introduction

Lakes sustain human life and economy but face severe ecological degradation from human activities and climate change [1,2,3,4]. As a vital component of lake ecosystems, the growth status of aquatic plants serves as an indicator of the ecological health of lakes. In addition, aquatic plants help regulate water quality and maintain the balance of the aquatic environment. They can adsorb nutrients and heavy metals from the water, thus aiding in purification [5]. Additionally, by absorbing carbon dioxide and releasing oxygen, they improve the physico-chemical environment and help maintain ecological balance in aquatic systems [6]. Furthermore, aquatic plants provide essential habitats and food sources for birds, fish, and other organisms, playing a vital role in preserving biodiversity [7]. Nevertheless, overgrowth of aquatic plants can be detrimental: decaying plant matter can cause secondary pollution, while residues that resist decomposition can accelerate sediment accumulation and promote marsh formation in lakes [7].
Based on their morphological and growth characteristics, aquatic plants are generally classified into three life forms: submerged, floating-leaved, and emergent plants. Submerged plants serve as habitats for zooplankton, benthic organisms, and fish, while playing an important role in the absorption and decomposition of nutrients, as well as in the concentration and accumulation of heavy metals [8]. Floating-leaved species help regulate phytoplankton growth and enhance water clarity [9,10]. Emergent plants provides nesting and refuge sites for birds and fish, and nearshore emergent communities help buffer water flow and attenuate wind-induced waves [9,10].
Given that different life forms of aquatic plants play distinct roles within aquatic ecosystems, it is essential to classify and extract aquatic vegetation accordingly.
Conventional monitoring approaches for aquatic vegetation—such as field sampling—yield accurate results but demand substantial time, labor, and material resources, and they are often incapable of capturing the continuous spatial distribution of vegetation [11]. In contrast, remote sensing techniques offer rapid and cost-effective monitoring capabilities, enabling the acquisition of large-scale spatial information. Moreover, the traceability of remote sensing imagery provides a solid foundation for long-term monitoring of aquatic vegetation.
The classification of aquatic vegetation primarily relies on three types of features: spectral characteristics, spectral indices derived from them, and texture features. Emergent and most floating-leaved vegetation display typical vegetation spectral profiles, with canopy spectra mainly influenced by factors such as canopy coverage, structural attributes, and biochemical parameters. Their spectral signatures are characterized by low reflectance in the blue and red regions and high reflectance in the green and near-infrared regions [11,12]. In contrast, the spectral behavior of submerged vegetation is additionally affected by aquatic environmental factors, including water transparency, depth, chlorophyll-a concentration, and suspended sediment concentration. As a result, submerged vegetation generally exhibits lower reflectance in the visible bands compared to emergent vegetation, floating-leaved vegetation [11,12]. All three aquatic vegetation life forms display pronounced surface roughness and textural characteristics.
Researchers commonly classify aquatic vegetation using spectral characteristics and spectral indices, applying both traditional approaches (e.g., decision trees) and modern machine learning algorithms (e.g., support vector machines and random forests). The decision tree method, known for its simplicity, efficiency, and interpretability, remains one of the most commonly used techniques in aquatic vegetation classification [13,14,15,16,17,18,19,20,21,22,23]. In this approach, besides the selection of suitable classification features, the determination of segmentation thresholds plays a crucial role in influencing the classification performance. Traditionally, optimal segmentation thresholds are determined through visual interpretation for each image. However, this manual process introduces subjectivity and becomes time-consuming for long-term monitoring. To mitigate these limitations, fixed-threshold methods have been applied by several studies [13,14,15,19,20,23]. Recognizing that the optimal threshold can vary depending on factors such as acquisition time, viewing angle, and atmospheric conditions, other researchers have explored automated threshold determination techniques [16,17,18,21,22]. With the rapid progress of artificial intelligence, modern machine learning methods have become increasingly prevalent in aquatic vegetation classification. Compared with traditional rule-based classification approaches, machine learning methods can automatically capture complex nonlinear relationships between features and target classes from large training datasets, leading to substantial improvements in both classification accuracy and generalization performance. Approaches such as support vector machines [24], random forests [25,26,27], and deep learning [28] have demonstrated significant potential. As computational power and remote sensing data availability continue to expand, deep learning has emerged as a powerful tool in this field. Compared with other approaches, deep learning can autonomously extract intricate spectral–spatial features directly from raw imagery, exhibiting superior feature representation and classification performance even in complex environmental backgrounds. Gao et al. [28], for instance, developed a ResUNet-based model using Sentinel-2 Multispectral Imager (MSI) data from lakes across the Yangtze River Basin to classify aquatic vegetation with high precision.
Currently, most remote sensing studies on aquatic vegetation classification rely primarily on optical satellite imagery as data sources. Commonly used datasets include Sentinel-2 MSI with 10 m spatial resolution [18,19,20,26,27,28], Gaofen-1 (GF-1) Wide Field of View (WFV) imagery with 16 m resolution [15], the Landsat series at 30 m resolution [13,14,21,22,25], Huanjing-1 (HJ-1) Charge-Coupled Device (CCD) data at 30 m resolution [16], Moderate-resolution Imaging Spectroradiometer (MODIS) data at 500 m resolution [17], and hyperspectral imagery [24]. In addition, some researchers have explored the integration of optical data and Synthetic Aperture Radar (SAR)—such as Sentinel-1 SAR—to improve the accuracy of aquatic vegetation classification [23,29,30]. Nevertheless, medium- to high-resolution multispectral satellites like Sentinel-2 MSI and the Landsat series typically have revisit periods of at least five days. During the peak growing season of aquatic vegetation in the Yangtze River Basin—when clear-sky conditions are scarce—it is often challenging to acquire valid imagery. Although SAR data such as Sentinel-1 can penetrate clouds, it cannot effectively penetrate the water surface, limiting its ability to capture information on submerged vegetation. To address the challenge that existing medium- to high-resolution optical satellites have relatively long revisit cycles—making it difficult to obtain sufficient cloud-free imagery during the peak growing season of aquatic vegetation—this study aims to develop a high-temporal-resolution dataset (effective revisit interval of about one day) by integrating imagery from China’s Gaofen-1/6 (GF-1/6) Wide Field of View (WFV) sensors and Huanjing-2A/B (HJ-2A/B) Charge-Coupled Device (CCD) instruments. Building on this dataset, we leverage the strong spectral–spatial feature learning capability of deep learning algorithms to investigate its potential for high-precision aquatic vegetation classification.

2. Materials

2.1. Study Area

This study selected six representative lakes within China’s Yangtze River Basin for the purpose of sampling (Figure 1), including Lake Taihu, Lake Chaohu, Lake Honghu, Lake Dianchi, Lake Dianshanhu, and Lake Caohai. These lakes are evenly distributed along the main course of the Yangtze River, spanning its upper (Lakes Dianchi and Caohai), middle (Lake Honghu), and lower reaches (Lakes Chaohu, Taihu and Dianshanhu), thus encompassing the basin’s major hydrological segments. The selected lakes exhibit both representativeness and diversity in the composition of aquatic vegetation life forms. As shown in Table 1, each lake supports at least two of the three dominant life forms—submerged, floating-leaved, and emergent plants—providing a comprehensive representation of aquatic vegetation growth conditions across the Yangtze River Basin [31]. The region lies within the subtropical monsoon climate zone, where rainfall and warmth occur concurrently [32]. The growing season of aquatic vegetation typically spans from April to October, with peak biomass observed during August and September [27,33,34,35]. During this period, the area is predominantly influenced by monsoonal circulation, resulting in frequent overcast and rainy weather [36,37].
Lake Taihu (30°55′N–31°34′N, 119°53′E–120°36′E), China’s third-largest freshwater lake, is located at the center of the Yangtze River Delta. The basin is one of the country’s most economically developed regions, with its population in 2024 accounting for 4.9% of China’s total and its GDP contributing 10.1% to the national economy [38], reflecting the intensity of human activities. The region experiences a subtropical monsoon climate with distinct seasons, abundant warmth, and plentiful rainfall. The long-term mean temperature ranges from 15 to 17 °C, and the average annual precipitation is approximately 1177 mm. Lake Taihu supports three primary life forms of aquatic vegetation—submerged, floating-leaved, and emergent species—showing a spatial pattern of sparse growth in the northern and western zones, while being widely distributed across the northeastern, eastern, and southern parts of the lake [39]. Given its large surface area, complex aquatic vegetation community, and the coexistence of intensive human activities with relatively complete lake ecosystems, Lake Taihu serves as a representative site for exploring the interactions between anthropogenic disturbances and aquatic vegetation dynamics in the Yangtze River Basin.

2.2. Image Data

The remote sensing data utilized in this study primarily consist of images from the GF-1/6 WFV and HJ-2A/B CCD satellites. GF-1, launched in April 2013, and GF-6, launched in June 2018, are both equipped with WFV sensors offering a spatial resolution of 16 m. Each satellite has an individual revisit period of 4 days, which shortens to 2 days when operating in tandem. Similarly, the HJ-2A and HJ-2B satellites, launched in September 2020, carry CCD sensors with multispectral bands of the same 16 m spatial resolution and a 4-day revisit cycle, reduced to 2 days in dual-satellite operation. When these four satellites function together as a complete observation network, the theoretical revisit frequency can be further improved to once per day. Taking Lake Taihu from April to October 2024 as an example (Figure 2). Over a total observation period of 214 days, complete satellite coverage was achieved on 173 days, of which 21 days provided high-quality imagery characterized by minimal cloud cover and sun glints.
The spectral bands for each sensor are listed in Table 2. This study mainly utilized the common spectral bands shared by all sensors—the blue (B1), green (B2), red (B3), and near-infrared (B4) bands—for analytical purposes. All satellite imagery used in this research is available from the China Centre for Resources Satellite Data and Application (CRESDA) (https://data.cresda.cn/#/home, accessed on 1 April 2025).

2.3. Field Validation Data

The validation areas for aquatic vegetation distribution were delineated through a comprehensive analysis combining historical literature and remote sensing interpretation results, with survey timing adjusted according to weather conditions and satellite overpass schedules. A standardized quadrat sampling method was employed to validate the spatial distribution patterns of aquatic vegetation in the field. Considering both the spatial resolution of the satellite imagery and the operational feasibility of fieldwork, each quadrat was uniformly designed as a 10 × 10 m2. By integrating ground-based observations with aerial imagery acquired via unmanned aerial vehicles (UAVs), the study conducted a systematic investigation and quantitative evaluation of aquatic vegetation community types and their coverage across the study area.
Between August 2023 and October 2024, eight systematic field validation surveys were conducted across all lakes within the study area. A total of 121 sampling points of aquatic vegetation distribution were obtained, comprising 38 sites of submerged vegetation, 9 of floating-leaved vegetation, 40 of emergent vegetation, and 34 of background categories. The spatial distribution of these validation sites is shown in Figure 3.

2.4. Measured Spectral Data

In this study, an Analytical Spectral Devices (ASD) FieldSpec spectroradiometer (Malvern Panalytical, Malvern, United Kingdom) was used to measure the spectral characteristics of aquatic vegetation and water bodies. Given that different life forms of aquatic vegetation vary in their vertical positioning relative to the water surface, resulting in distinct spectral features, a differentiated measurement strategy was applied. For submerged vegetation and open water, above-water spectral measurements were performed to effectively separate water-leaving radiance from atmospheric scattering signals [40], whereas for floating-leaved and emergent vegetation, conventional field spectral measurement methods were used to directly obtain canopy reflectance.
Field validation and supplementary water body experiments were conducted under clear and cloud-free conditions. A total of 78 reflectance spectra were obtained, representing four categories: submerged vegetation (n = 13), floating-leaved vegetation (n = 10), emergent vegetation (n = 25) and open water (n = 30) (Figure 4a–d). The spectral data of various surface features were averaged, together with algal bloom spectra, used to generate characteristic reflectance curves for different surface types (Figure 4e). Results indicate that in the visible wavelength range, water shows slightly higher reflectance than other surfaces, with submerged vegetation exhibiting the lowest reflectance. In the near-infrared range, emergent vegetation demonstrate the typical high-reflectance pattern of green vegetation, greatly exceeding that of floating-leaved vegetation, algal blooms, and submerged vegetation, while water remains the darkest target.

3. Methods

3.1. Image Preprocessing

Using Sentinel-2 MSI orthorectified imagery provided by European Space Agency (ESA) (https://browser.dataspace.copernicus.eu/, accessed on 1 April 2025, 10 m spatial resolution), together with SRTM DEM data (https://dwtkns.com/srtm30m/, accessed on 1 April 2025, 30 m spatial resolution), we refined the original RPC model of the domestic satellites and applied the corrected RPC model to geometrically rectify the imagery [41,42,43,44].
Radiometric calibration was applied to the geometrically corrected images using the official calibration coefficients released by CRESDA (https://mp.weixin.qq.com/s/uREY-V33lQTPqNlCtKuuww, accessed on 1 April 2025), converting raw DN values into spectral radiance. For historical images, absolute radiometric calibration coefficients published within six months before or after the imaging date were adopted. For recent images, if coefficients within that six-month window were unavailable, the latest officially released values were used. This approach ensured that the interval between image acquisition and coefficient generation never exceeded 15 months, thereby maintaining the reliability of radiometric calibration.
Finally, atmospheric correction was performed using radiative transfer models such as 6S [45,46,47,48] or MODTRAN [49,50], converting top-of-atmosphere radiance into surface reflectance. The models simulate atmospheric scattering and absorption processes while incorporating site-specific atmospheric parameters at the time of acquisition—such as aerosol type and water vapor content—to effectively eliminate atmospheric effects and improve spectral fidelity. Based on the study area’s geographic setting, imaging date, and data quality, the atmospheric model was configured as mid-latitude summer, the aerosol model as rural, and visibility was set to 23 km, ensuring high-accuracy atmospheric correction.

3.2. Equivalent Reflectance Computation

Using measured spectral data together with the spectral response function (SRF) provided by CRESDA (https://mp.weixin.qq.com/s/uREY-V33lQTPqNlCtKuuww, accessed on 1 April 2025), the equivalent surface reflectance of each band was derived through convolution integration [51]. Given the differences in spectral measurement approaches for various surface types, appropriate equivalent reflectance computation methods were adopted for each category.
For submerged vegetation and open-water surfaces, the equivalent water remote sensing reflectance ( R rs ) was derived from field measurements and subsequently multiplied by π to obtain the equivalent surface reflectance (R) [52]. The corresponding computation was expressed as
R rs ( b a n d i ) = λ 1 λ 2 R rs ( λ ) S R F ( λ ) d ( λ ) λ 1 λ 2 S R F ( λ ) d ( λ ) , R ( b a n d i ) = R rs ( b a n d i ) × π .
where R rs ( b a n d i ) denotes the equivalent water remote-sensing reflectance for the i-th spectral band ( sr 1 ), R rs ( λ ) is the measured water remote sensing reflectance ( sr 1 ), and λ 1 to λ 2 denote the wavelength range of the band.
The equivalent surface reflectance of floating-leaved and emergent vegetation was calculated using the following formula:
R ( b a n d i ) = λ 1 λ 2 R ( λ ) S R F ( λ ) d ( λ ) λ 1 λ 2 S R F ( λ ) d ( λ )
where R ( b a n d i ) represents the equivalent surface reflectance for the i-th spectral band, while R ( λ ) refers to the measured surface reflectance, and λ 1 to λ 2 denote the wavelength range of the band.

3.3. Remote Sensing Interpretation Features

This study differentiated various surface types primarily by analyzing their spectral and textural characteristics, as summarized in Table 3.

3.4. Sample Preparation

The study defined the lake boundary as the spatial reference for sample generation. Remote sensing imagery containing only the near-infrared, red, and green bands was cropped to produce raster image datasets, and the corresponding classification vector data were converted into label rasters of identical spatial extent. Using a fixed-size sliding window with a 10% overlap and zero-padding along the boundaries, spatially aligned image–label pairs of 256 × 256 pixels were generated. To address the class imbalance caused by the dominance of background pixels such as open water and algal bloom, all pairs consisting solely of background classes were excluded, resulting in 836 valid sample pairs. Given the intensive workload involved in manual interpretation of aquatic vegetation types, data augmentation techniques were employed to enhance dataset diversity and utilization. Augmentation methods included rotations of 90° and 270°, horizontal and vertical flips, and diagonal mirroring. After augmentation, the dataset expanded to 5016 pairs, which were subsequently divided into training and validation sets at a ratio of 8:2.

3.5. Network Model Architecture

This study employed an advanced encoder–decoder segmentation framework, namely U-Net++ integrated with an EfficientNet-B5 backbone. The encoder was initialized with weights pretrained on the ImageNet dataset, leveraging transfer learning to enhance the efficiency and robustness of multi-scale feature extraction—from high-level semantic representations to fine-grained textural details—across multi-channel remote sensing imagery. The decoder adopted the U-Net++ architecture, an enhanced version of the classic U-Net, characterized by its densely nested skip connections that allow more precise fusion of hierarchical features captured at different encoding stages [53]. By combining EfficientNet’s powerful feature representation with U-Net++’s superior capability in reconstructing spatial details and object boundaries, this hybrid architecture provided a robust foundation for the precise segmentation of aquatic vegetation with complex morphological structures.

3.6. Experimental Settings

All experiments were implemented on an NVIDIA GeForce RTX 4080 SUPER GPU equipped with 16 GB of memory. To balance model performance, contextual feature capture, and computational efficiency, the input image size was normalized to 256 × 256 pixels, with a batch size of 4. The segmentation framework combined U-Net++ and an EfficientNet-B5 backbone, optimized for a four-class classification task. Parameter updates were carried out using the Adam optimizer, with an initial learning rate of 1 × 10 4 that decayed progressively to 1 × 10 6 following a cosine annealing schedule.
A hybrid loss function integrating Dice Loss and Focal Loss was designed to improve recognition of underrepresented classes. To mitigate overfitting, training was run for 200 epochs with an early stopping criterion—halting automatically if validation performance showed no improvement for 20 consecutive epochs. Validation was executed after each epoch, and model checkpoints were saved every five epochs to ensure training stability and traceability.

3.7. Model Evaluation

3.7.1. Source of Evaluation Data

To thoroughly evaluate the model’s classification performance, this study utilized multi-source validation datasets. The visual interpretation results derived from remote sensing imagery provided polygonal reference data representing the distribution of aquatic vegetation, while field validation provided point-based ground truth samples. Together, these datasets formed a comprehensive, multi-scale validation framework. This integrative approach allowed simultaneous evaluation of the model’s capability to capture large-scale spatial patterns and its performance in localized classifications, ensuring a systematic and robust assessment of overall model performance.

3.7.2. Evaluation Method Design

Surface reference data were generated from Sentinel-2 MSI imagery acquired on the same date as the target image with a higher spatial resolution of 10 m. Using the spectral, textural, and spatial distribution features of aquatic vegetation, manual visual interpretation was performed to delineate high-accuracy reference polygons. Within the study region, we selected three representative lakes for model evaluation: Lake Taihu, the largest lake with extensive and highly diverse aquatic vegetation; Lake Chaohu, the second largest but characterized by sparse vegetation and relatively few life forms; and Lake Caohai, the smallest lake, where vegetation is widely distributed but exhibits low diversity.
A point-based validation dataset was established by integrating field validation data collected under favorable weather conditions with temporally and spatially matched remote sensing imagery. Given the temporal stability of aquatic vegetation spectral characteristics [33,35] and the scarcity of clear-sky days that constrained the acquisition of high-quality imagery, we selected high-quality images acquired within 15 days of the field validation and excluded from model training as the matching data source. Due to field access limitations, boat-based observations were restricted to the marginal zones of aquatic vegetation, making it challenging to capture pure vegetation pixels. Thus, a coverage threshold of 50% was defined as the minimum level of aquatic vegetation that can be reliably detected through remote sensing. Furthermore, to account for GPS positioning errors, vessel drift caused by wind and water currents, and to minimize uncertainties in spatial distribution and coverage due to temporal differences between imagery and field data, a buffer zone with a 50 m radius (approximately three pixels) centered on each sampling point was established. Classification results were deemed correct when at least one classified pixel within the buffer zone matched the corresponding field observation.

3.7.3. Evaluation Metrics

A pixel-level, dual-dimensional assessment framework was employed in this study. Class-level evaluation metrics were used to quantify the model’s capability in distinguishing individual categories, while overall metrics was applied to evaluate its global classification performance. This integrated approach enabled a comprehensive evaluation of the model from both local and overall perspectives.
(1)
Class Evaluation
The classification performance for each category was assessed using standard evaluation metrics, including Intersection over Union (IoU), Precision, Recall, and F1-Score. The corresponding formulas are presented as
I o U = T P T P + F P + F N
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
F 1 S c o r e = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l = 2 × T P 2 × T P + F P + F N
where T P represents the number of correctly classified positive pixels, F P represents the number of negative pixels misclassified as positive, and F N represents the number of positive pixels misclassified as negative.
(2)
Overall Evaluation
The overall classification performance was evaluated using four standard metrics: mean Intersection over Union (mIoU), mean Pixel Accuracy (mPA), Overall Accuracy (OA), and the Kappa coefficient.
The mIoU metric quantifies the overlap between predicted and reference regions, and is calculated as
m I o U = 1 n × i = 1 n I o U i
where I o U i denotes the Intersection over Union for the i-th category, and n represents the total number of categories.
Overall Accuracy (OA) evaluates the model’s global classification correctness across all pixels, serving as an indicator of its overall performance. Mean Pixel Accuracy (mPA), on the other hand, assesses the model’s per-class pixel-level accuracy, providing insight into its ability to distinguish minor or less represented classes. The corresponding formulas are expressed as
O A = i = 1 n T P i N
m P A = 1 n × i = 1 n T P i T P i + F N i
where T P i represents the number of correctly classified pixels for the i-th class, and F N i denotes the number of pixels from the i-th class that are incorrectly assigned to other categories, and N refers to the total number of pixels.
The Kappa coefficient quantifies the agreement between the model’s classification results and the ground truth, accounting for the possibility of random agreement. Its calculation is expressed as
K a p p a = p 0 p e 1 p e
where p 0 denotes the ratio of correctly classified pixels to the total number of pixels, corresponding to the Overall Accuracy (OA), and p e is defined as follows:
p e = i = 1 n ( T P i + F N i ) × ( T P i + F P i ) N × N
where F P i denotes the number of pixels from other categories that are incorrectly assigned to the i-th class.

3.8. Statistical Approaches and Error Analysis Techniques

Pearson’s correlation coefficient was used to quantify the relationship between the two datasets: image reflectance versus equivalent reflectance in the atmospheric correction assessment, and inter-sensor band-level reflectance consistency in the sensor consistency evaluation.
Root Mean Square Error (RMSE) was further applied to measure their quantitative differences, reflecting deviations from equivalent reflectance in the atmospheric correction analysis and discrepancies in band reflectance between the two sensors in the consistency assessment.
R M S E = i = 1 n ( X obs , i X pred , i ) 2 n
where X obs , i represents observed data, and X pred , i denotes predicted data, n stands for total number of observations.

4. Results

4.1. Atmospheric Correction Evaluation

Given the strong spatiotemporal variability of water optical properties and the comparatively stable spectral characteristics of aquatic vegetation, this study employed a data collection strategy integrating both synchronous and quasi-synchronous observations to assess atmospheric correction performance. For water bodies, strictly synchronous matching was performed between in situ spectra and satellite image data acquired on the same day. For aquatic vegetation, quasi-synchronous matching was conducted using spectra and imagery collected within 7 days, ensuring consistent surface cover types for the matched pixels. During data processing, the equivalent surface reflectance was derived from field spectra using SRF and quantitatively compared with the surface reflectance retrieved from corresponding image pixels. As illustrated in Figure 5, the validation results showed that most scatter points followed a well-aligned 1:1 linear relationship, with Pearson’s r consistently exceeding 0.83 (p < 0.001) and RMSE remaining below 0.115. These findings confirm that the atmospheric correction approach employed in this study provides high accuracy and robustness.

4.2. Sensor Consistency Evaluation

In this study, Chinese domestic multispectral satellite data were jointly employed for model development, incorporating imagery from multiple sensors, including GF-1 WFV (four cameras), GF-6 WFV (one camera), and HJ-2A/B CCD (four cameras). As shown in Figure 6, inherent differences among the Spectral Response Functions (SRFs) of these sensors can introduce systematic deviations in surface reflectance products. To ensure the effective integration of multisensor datasets within the deep learning framework, we conducted a twofold evaluation of spectral consistency: (1) simulating and comparing sensor-specific reflectance using in situ surface spectra combined with each sensor’s SRF; and (2) directly comparing surface reflectance products derived from quasi-synchronous imagery. This comprehensive assessment enabled a quantitative analysis of the spectral discrepancies among reflectance products generated by different sensors.

4.2.1. Sensor Consistency Evaluation Based on Spectral Response Function

For multi-camera mosaic sensors, the equivalent surface reflectance of each camera was first derived using the in situ spectra and the respective SRF of each camera. The equivalent reflectances from all cameras within a single sensor were then averaged to represent the overall equivalent surface reflectance of that sensor. Cross-validation of band-matched equivalent reflectance across different sensors (Figure 7) demonstrates that scatter points for all bands align closely with the 1:1 reference line. Pearson’s r exceeds 0.81 for every band (p < 0.001), and RMSE remains below 0.012, indicating strong spectral consistency among sensors.

4.2.2. Sensor Consistency Evaluation Based on Quasi-Synchronous Imagery

Due to the scarcity of clear-sky conditions in the study area, it is difficult to obtain perfectly synchronous multi-source satellite data. To address this, we selected a pair of quasi-synchronous images—a GF-1 WFV scene acquired on 23 August, 2024, and an HJ-2B CCD scene captured one day later—to evaluate sensor consistency at the surface reflectance level. To precisely assess the spectral consistency of specific surface categories, the imagery was overlaid with a pre-established vector dataset representing aquatic vegetation classifications for comparative analysis. Due to the highly variable optical characteristics of water bodies, their spectral signatures can change noticeably even within a single day, while the spatial distribution of algal blooms also shifts over time. Consequently, this study excluded background pixels such as open water and algal bloom areas, concentrating instead on comparing the reflectance of the three primary aquatic vegetation types—submerged, floating-leaved, and emergent vegetation. As shown in Figure 8, the reflectance scatter points of all three aquatic vegetation life forms cluster closely around the 1:1 reference line across all spectral bands. Pearson’s r exceeds 0.65 for each band (p < 0.001), and the RMSE remains below 0.040, demonstrating strong spectral consistency between the two sensor datasets.
Combining the results from SRF simulations and quasi-synchronous imagery cross-validation, this study demonstrates that the four selected sensors maintain strong spectral consistency in surface reflectance products. Therefore, additional cross-sensor normalization or calibration is unnecessary. This finding suggests that these heterogeneous datasets are suitable for collaborative classification and can provide a solid data basis for building a unified joint model.

4.3. Selection of Input Parameters and Model Components

In remote sensing image semantic segmentation, the selection of input parameters can strongly influence model performance. To verify the rationality and necessity of the input configurations used in this study, comparative experiments were conducted from two aspects: input band combinations and input block size. All experiments were performed under the same training strategy and network architecture, with each configuration tested in three repeated trials.
First, in terms of input bands, we tested three configurations: using only near-infrared, red, and green (NIR + R + G); adding the blue band; and adding the normalized difference vegetation index (NDVI). As shown in Table 4, the NIR + R + G configuration achieved the most favorable performance across three trials, outperforming the other setups. This indicates that including additional bands such as blue or NDVI can introduce redundancy, potentially increasing the model’s learning complexity and reducing its generalization capability.
Secondly, in terms of input block size, some lakes are relatively small, with dimensions less than 512 pixels, making the use of 512 × 512 image blocks for training and prediction prone to errors. Therefore, we focused on testing 256 × 256 and 128 × 128 block sizes. The results indicated that the 256 × 256 size provides an optimal balance between performance and computational efficiency, while smaller blocks ( 128 × 128 ) suffer from lower performance due to limited spatial context.
In conclusion, the experiments demonstrate that using the NIR + R + G band combination along with a 256 × 256 input block size represents the optimal configuration for this task. This setup achieves a balance between computational efficiency and enhanced model performance and robustness, offering a solid experimental basis for further model comparisons and practical applications.
Based on the identified optimal input configuration (band combination and input block size), a series of comparative experiments were designed to systematically evaluate the effectiveness of each component in the final segmentation model and its contribution to overall performance.
Specifically, key modules of the model architecture, including the encoder, decoder, and loss function, were individually tested by removing or substituting them to quantify performance changes. As illustrated in Table 5, these experiments provided insights into how each component contributes to performance improvement and confirmed the rationality and necessity of the modeling approach proposed in this study.

4.4. Model Performance

The model performance stabilized after around ten training epochs, where the training loss converged to 0.066, the validation loss to 0.086, and the training mIoU surpassed 90%. As shown in Figure 9, without transfer learning, both the validation loss and training mIoU fluctuated noticeably and failed to converge effectively. Ultimately, the optimal model demonstrated robust performance on the validation set, with an mIoU of 90.16%, an mPA of 95.27%, an OA of 99.11%, and a Kappa coefficient of 0.94.
To comprehensively assess the model’s classification performance, we established a multi-source validation framework. At the macro level, polygonal ground-truth data derived from visual interpretation were used to evaluate the spatial consistency of classification results. At the micro level, point-based field validation data were employed to examine the model’s local classification accuracy. This framework enables a multidimensional and systematic evaluation of the model’s classification capabilities.

4.4.1. Model Evaluation Based on Visual Interpretation

In this study, visually interpreted results from high-resolution Sentinel-2 MSI imagery (10 m spatial resolution) were used as benchmark ground truth to evaluate the model’s classification performance across a broad regional scale. Evaluation was conducted using a fully independent test dataset (Lake Taihu imagery acquired on 16 August 2020; Lake Chaohu on 12 August 2023; and Lake Caohai on 18 October 2024), which was strictly excluded from model training. As shown in Figure 10, the model achieved a relatively good overall performance, with mIoU = 79.10%, mPA = 86.42%, OA = 98.47%, and a Kappa coefficient of 0.86. At the class level, the IoU, Precision, Recall, and F1-Score for all aquatic vegetation life forms reached 61.13%, 77.19%, 74.61%, and 75.88%, respectively. Some confusion occurred between floating-leaved and emergent vegetation due to their similar spectral and textural characteristics, resulting in slightly lower classification performance for these types. Overall, the quantitative assessment confirms that the proposed deep learning model effectively captures the spatial distribution patterns of different aquatic vegetation life forms, highlighting its strong potential for large-scale aquatic vegetation monitoring via remote sensing.

4.4.2. Model Evaluation Based on Field Validation

Field validation data were used to evaluate the model’s classification performance in representative areas. Using field validation data collected under favorable weather conditions and remote sensing images obtained within a 15-day interval, a benchmark dataset comprising 101 field samples was established. These included 31 samples of submerged vegetation, 7 of floating-leaved vegetation, 29 of emergent vegetation, and 34 of background classes. As shown in Figure 11, the model achieved producer’s accuracies (PA) of 51.61% for submerged vegetation, 85.71% for floating-leaved vegetation, and 68.97% for emergent vegetation, while the user’s accuracies (UA) were 100%, 75.00%, and 90.91%, respectively. The OA reached 75.25%, with a Kappa coefficient of 0.65, indicating that the model maintains reliable classification performance in field-validated regions.

4.5. Comparison with Other Methods

We further compared our model with other widely adopted deep learning semantic segmentation frameworks. DeepLabv3+ excels at capturing multi-scale contextual information, making it particularly effective for large-scale and complex landscapes with diverse object sizes, while HRNet demonstrates superior performance in delineating fine boundaries and small-scale features due to its high-resolution feature maintenance mechanism. We performed three independent experimental runs for each algorithm. As illustrated in Figure 12, the model we trained achieves marginally higher mIoU and OA values compared with the other two models.

4.6. Spatiotemporal Variation of Aquatic Vegetation in Lake Taihu

The trained deep learning model was employed to classify and extract remote sensing imagery of Lake Taihu captured during the peak growing season of aquatic vegetation over the past five years. Based on these classification results, the overall growth conditions of aquatic vegetation in Lake Taihu were analyzed. As illustrated in Figure 13, aquatic vegetation in Lake Taihu is primarily concentrated in the eastern portion of the lake. Submerged vegetation dominates the central region, while floating-leaved and emergent vegetation are mainly found in the southern area. These spatial patterns align well with previously reported field investigations and remote sensing observations [27,34,39]. The total area of aquatic vegetation in Lake Taihu peaked in 2024 at approximately 189.85 km 2 , then declined to its lowest level in 2025, about 114.53 km 2 .
Based on multi-temporal satellite imagery from the growing seasons of 2024 and 2025—the years with the largest and smallest aquatic vegetation coverage in the past five years—aquatic vegetation was extracted and analyzed. The mean monthly aquatic vegetation area was used to characterize vegetation growth conditions in each month. As illustrated in Figure 14, aquatic vegetation is in its germination stage in April, enters a rapid growth phase from May to June, stabilizes gradually from July onward, and peaks around September.
Relative to other years, the total aquatic vegetation coverage in Lake Taihu in 2025 declined markedly, primarily due to the drastic degradation of submerged vegetation. A comparison of the monthly variations between 2024 and 2025 showed that the submerged vegetation area in 2025 had already fallen well below the previous year’s level by June and remained extremely low thereafter. To mitigate this decline, management agencies could strengthen water quality and water level regulation, enhance ecological monitoring, and conduct timely artificial replanting to restore submerged vegetation, thereby reducing algal blooms and curbing further eutrophication.

5. Discussion

In this study, we integrate domestic multispectral satellite observations (GF-1/6 WFV and HJ-2A/B CCD) with a deep learning classification framework to enable detailed mapping of lake aquatic vegetation across the Yangtze River Basin. Although deep learning has proven highly effective for automatically learning spatial–spectral representations and reducing reliance on handcrafted features, its use in this region has long been constrained by limited data sources and low observation frequency. For instance, despite the high spectral resolution and mature product system of Sentinel-2 MSI, persistent cloudy and rainy conditions during the peak growing season of aquatic vegetation (July–September) greatly restrict the availability of clear-sky imagery. To address this challenge, we incorporate multi-source domestic satellite imagery, reducing the effective revisit interval to approximately 1 day and substantially increasing the number of usable scenes and the timeliness of monitoring. Furthermore, by integrating U-Net++ with EfficientNet-B5, we develop a high-accuracy classification model whose mIoU and Kappa scores improve by 11% and 0.07, respectively, compared with previous work [28]. This framework provides a stronger data foundation and technical pathway for large-scale, high-frequency, and long-term remote sensing monitoring of aquatic vegetation.
Although our sample set includes several representative lakes—such as Lake Taihu, Lake Chaohu, and Lake Honghu—spanning diverse vegetation types and markedly different geographic settings, its spatial coverage is still limited. When the model is transferred to other lakes, variations in spectral signatures and spatial structural characteristics may reduce its accuracy in fine-grained classes, even if overall classification performance remains high. Moreover, the images used for sample construction were primarily acquired under favorable conditions with low cloud cover and minimal sun-glint contamination, while clouds and glint are common in real-world operational scenarios. Clouds obscure surface information outright, and sun glint makes it difficult for the model to recover the true characteristics of targets such as submerged vegetation. This often leads to missing or unretrievable vegetation patches—an inherent limitation of optical remote sensing in complex environments. Additionally, submerged vegetation is highly sensitive to water transparency, depth, chlorophyll-a concentration, and suspended matter levels [11,12], frequently resulting in situations where vegetation is clearly visible in the field but indistinguishable in satellite imagery. Consequently, even if evaluations based on visual interpretation appear promising, model performance may still be weaker when validated against in situ measurements.
To overcome these limitations, future research could improve usability and generalization along two directions. First, by taking advantage of the seasonal growth dynamics and relatively stable spatial patterns of aquatic vegetation, approaches incorporating spatial geometric constraints, neighborhood-based spatial inference, or temporal interpolation models could be employed to estimate vegetation conditions in areas obscured by dense clouds or strong sun glint. Such methods would help reduce spatiotemporal data gaps and enhance the completeness of monitoring results. Second, although domestic satellite constellations substantially increase the availability of optical imagery, some years or lake regions may still experience several consecutive weeks without usable scenes due to prolonged cloudy and rainy weather. To improve the continuity and robustness of long-term time-series monitoring, integrating Sentinel-1 or other SAR datasets as complementary sources would enable all-weather aquatic-vegetation observation through optical–radar data fusion, providing more reliable temporal information under challenging meteorological conditions.
In conclusion, the aquatic-vegetation monitoring framework developed in this study—combining Chinese domestic satellite imagery with deep learning—offers notable enhancements in image accessibility, model robustness, and the spatiotemporal continuity of monitoring. Moreover, it establishes a strong data and technical foundation for future studies on vegetation growth dynamics, exploration of ecological drivers, and the conservation of lake ecosystems in the Yangtze River Basin.

6. Conclusions

Based on 16 m spatial resolution domestic satellite data (GF-1/6 WFV and HJ-2A/B CCD), this study constructed aquatic vegetation sample sets for representative lakes in the Yangtze River Basin using visual interpretation. The datasets were split into training and validation sets at an 8:2 ratio. The U-Net++/EfficientNet-B5 model trained on the training set achieved high performance on the validation set, with mIoU = 90.16% and mPA = 95.27%. On independent test data derived from higher-resolution imagery (Sentinel-2 MSI) through visual interpretation, the model attained mIoU = 79.10% and mPA = 86.42%, with IoU and F1-Score for all aquatic vegetation life forms reaching 61.13% and 75.88%, respectively. Field-based evaluation yielded OA = 75.25% and Kappa = 0.65. When applied to peak-season images of Lake Taihu over the past five years, we found that aquatic vegetation in Lake Taihu is predominantly distributed in the eastern part of the lake, with total coverage peaking in 2024 and declining to its lowest in 2025, largely due to the drastic degradation of submerged vegetation.
The approach presented in this study offers reliable support for monitoring aquatic vegetation in lakes across the Yangtze River Basin. Our analysis demonstrates that the four sensors (GF-1/6 WFV and HJ-2A/B CCD) exhibit high consistency, satisfying the requirements for collaborative classification. Compared with other commonly used medium- to high-resolution satellites, these sensors provide substantially improved temporal coverage, providing more effective data during the peak growth period of aquatic vegetation when sunny days are limited. Furthermore, we identified that the sharp decline in Lake Taihu’s aquatic vegetation area in 2025 was primarily driven by the degradation of submerged vegetation, a trend that could be detected as early as June. This suggests that the technology developed in this study enables real-time monitoring of aquatic vegetation dynamics, facilitating early problem detection and timely management interventions.
It is important to note that, given the limited extent of the sample set, the model may show some inaccuracies when extended to other lakes, different river basins, or imagery impacted by clouds and complex weather conditions. Additionally, factors such as sun glint and water environment conditions can cause submerged vegetation to be visible in situ yet undetectable in remote sensing images, which can hinder the model’s ability to accurately extract these features.

Author Contributions

Conceptualization, Q.S.; methodology, Y.S.; software, Y.S.; validation, Y.S.; formal analysis, Q.S. and Y.Y.; investigation, Y.S.; resources, X.W., H.Z. (Huan Zhao), H.Z. (Haobin Zhang), H.G. and Y.Z.; data curation, H.G. and Y.Z.; writing—original draft preparation, Y.S.; writing—review and editing, Q.S. and Y.Y.; visualization, H.G.; supervision, Q.S. and Y.Y.; project administration, Q.S. and Y.Y.; funding acquisition, Z.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (General Program of National Natural Science Foundation of China, No. 41971381).

Data Availability Statement

The data presented in this study are available on request from the corresponding author due to privacy reason.

Acknowledgments

The authors extend their sincere gratitude to Qian Shen from the Key Laboratory of Digital Earth Science, Aerospace Information Research Institute, Chinese Academy of Sciences, for her assistance in writing. We are also deeply grateful to Qiang Gao, Pengfei Ji, Xin Zhou, Qian Wang, and Chiyi Yang for their contributions in preparing samples using visual interpretation. Additionally, we wish to thank the anonymous reviewers for their valuable and constructive comments on this study.

Conflicts of Interest

Autor Yuting Zhou was employed by Jiangsu Tianyan Environment Technology Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Nanjing Institute of Geography and Limnology, Chinese Academy of Sciences. Research Report on China’s Lake Ecosystem; Technical report; Nanjing Institute of Geography and Limnology, Chinese Academy of Sciences: Nanjing, China, 2022. [Google Scholar]
  2. Ma, R.; Yang, G.; Duan, H.; Jiang, J.; Wang, S.; Feng, X.; Li, A.; Kong, F.; Xue, B.; Wu, J.; et al. China’s Lakes at Present: Number, Area and Spatial Distribution. Sci. China Earth Sci. 2011, 54, 283–289. [Google Scholar] [CrossRef]
  3. Yang, G.; Ma, R.; Zhang, L.; Jiang, J.; Yao, S.; Zhang, M.; Zeng, H. Lake Status, Major Problems and Protection Strategy in China. J. Lake Sci. 2010, 22, 799–810. [Google Scholar]
  4. Zhang, Y.; Qin, B.; Zhu, G.; Song, C.; Deng, J.; Xue, B.; Gong, Z.; Wang, X.; Wu, J.; Shi, K.; et al. Importance and Main Ecological and Environmental Problems of Lakes in China. Chin. Sci. Bull. 2022, 67, 3503–3519. [Google Scholar] [CrossRef]
  5. Cao, C. Effects of Aquatic Vascular Plants on the Taihu Lake Ecosystem. Chin. J. Ecol. 1987, 6, 37–39+19. [Google Scholar]
  6. Maberly, S.C.; Gontero, B. Trade-Offs and Synergies in the Structural and Functional Characteristics of Leaves Photosynthesizing in Aquatic Environments. In The Leaf: A Platform for Performing Photosynthesis; Adams, W.W., III, Terashima, I., Eds.; Springer International Publishing: Cham, Switzerland, 2018; pp. 307–343. [Google Scholar] [CrossRef]
  7. Yang, Q. Ecological Functions of Aquatic Vegetation in East Taihu Lake and Its Reasonable Regulation. J. Lake Sci. 1998, 10, 67–72. [Google Scholar] [CrossRef][Green Version]
  8. Su, S.; Yao, W. A Brief Review on Mutual Relationship Between Submerged Macrophytes and Environment. J. Agro-Environ. Sci. 2002, 21, 570–573. [Google Scholar]
  9. Zhong, Y.; Hu, H.; Qian, Y. Advances in Utilization of Macrophytes in Water Pollution Control. Tech. Equip. Environ. Pollut. Control 2003, 4, 36–40. [Google Scholar]
  10. Ni, J.; Wang, W.; Xie, G.; Lian, H.; Wu, W.; Chen, L.; Sha, H. The Application Progress of Aquatic Plants in Water Ecological Restoration. Environ. Prot. Technol. 2016, 22, 43–47. [Google Scholar]
  11. Luo, J.; Yang, J.; Duan, H.; Lu, L.; Sun, Z.; Xin, Y. Research Progress of Aquatic Vegetation Remote Sensing in Shallow Lakes. J. Remote Sens. 2022, 26, 68–76. [Google Scholar]
  12. Xie, Y.; Li, J.; Li, J.; Zhang, F.; Shen, Q.; Wu, Y.; Zhang, B.; Ren, C. Progress and Prospect of Optical Remote Sensing Monitoring of Aquatic Vegetation in Lakes and Reservoirs. Environ. Monit. Forewarn. 2019, 11, 52–58. [Google Scholar]
  13. Cai, D.; Wei, W. Study on Remote Sensing Information Extraction of Aquatic Vegetation Based on Decision Tree. J. Anhui Agric. Sci. 2009, 37, 7615–7616. [Google Scholar] [CrossRef]
  14. Tao, T.; Ruan, R.; Zhang, L.; Fu, Q. Extraction Method of Wetland Vegetation Information in Hongze Lake. Geospat. Inf. 2017, 15, 93–96+99. [Google Scholar]
  15. Chen, Q.; Yu, R.; Hao, Y.; Wu, L.; Zhang, W.; Zhang, Q.; Bu, X. A New Method for Mapping Aquatic Vegetation Especially Underwater Vegetation in Lake Ulansuhai Using GF-1 Satellite Data. Remote Sens. 2018, 10, 1279. [Google Scholar] [CrossRef]
  16. Yan, D.; Zhou, X.; Liu, W.; Luo, J.; Rui, J.; Wang, Z.; Yu, Y. An Algorithm for Determining Remote Sensing Classification Threshold of Aquatic Vegetation Based on Gauss Fitting. J. Xi’an Univ. Sci. Technol. 2018, 38, 776–782. [Google Scholar] [CrossRef]
  17. Cao, P.; Liang, Q.; Li, S. A Novel Remote Sensing Simultaneous Monitoring Method for Cyanobacteria Blooms and Aquatic Vegetation in Taihu Lake Based on Otsu Algorithm. Jiangsu Agric. Sci. 2019, 47, 288–294. [Google Scholar] [CrossRef]
  18. Wang, Z.; Xin, C.; Sun, Z.; Luo, J.; Ma, R. Automatic Extraction Method of Aquatic Vegetation Types in Small Shallow Lakes Based on Sentinel-2 Data:A Case Study of Cuiping Lake. Remote Sens. Inf. 2019, 34, 132–141. [Google Scholar]
  19. Yang, J.; Luo, J.; Lu, L.; Sun, Z.; Cao, Z.; Zeng, Q.; Mao, Z. Changes in Aquatic Vegetation Communities Based on Satellite Images Before and After Pen Aquaculture Removal in East Lake Taihu. J. Lake Sci. 2021, 33, 507–517. [Google Scholar]
  20. Guo, H.; Han, X. Remote Sensing Extraction of Aquatic Vegetation Information inCoal Mining Subsidence Wetland Based on the Decision Tree Model. J. Anhui Univ. Sci. Technol. (Nat. Sci.) 2022, 42, 64–70. [Google Scholar]
  21. Chen, Y.; Zhu, Y. Remote Sensing Monitoring of Aquatic Vegetation Groups and Algal Blooms in Typical Inland Lakes in the Yellow River Basin. J. Irrig. Drain. 2022, 41, 81–88. [Google Scholar] [CrossRef]
  22. Luo, J.; Ni, G.; Zhang, Y.; Wang, K.; Shen, M.; Cao, Z.; Qi, T.; Xiao, Q.; Qiu, Y.; Cai, Y.; et al. A New Technique for Quantifying Algal Bloom, Floating/Emergent and Submerged Vegetation in Eutrophic Shallow Lakes Using Landsat Imagery. Remote Sens. Environ. 2023, 287, 113480. [Google Scholar] [CrossRef]
  23. Han, S.; Ruan, R.; Fu, Q.; Xu, H.; Heng, X. Extraction of aquatic vegetation in Hongze Lake National Wetland Park based on Sentinel-1 and Sentinel-2 images. J. Nanjing For. Univ. Nat. Sci. Ed. 2024, 48, 19–26. [Google Scholar]
  24. Pande-Chhetri, R.; Abd-Elrahman, A.; Jacoby, C. Classification of Submerged Aquatic Vegetation in Black River Using Hyperspectral Image Analysis. Geomatica 2014, 68, 169–182. [Google Scholar] [CrossRef]
  25. Shi, H.; Li, X.; Niu, Z.; Li, J.; Li, Y.; Li, N. Remote Sensing Information Extraction of Aquatic Vegetation in Lake Taihu Based on Random Forest Model. J. Lake Sci. 2016, 28, 635–644. [Google Scholar] [CrossRef]
  26. Zhang, P.; Zhang, F.; Li, J.; Xie, Y.; Zhang, B. Aquatic Vegetation Extraction of Yugiao Reservoir Based on Sentinel-2 Image Feature Optimization. Ecol. Sci. 2023, 42, 40–48. [Google Scholar] [CrossRef]
  27. Zhang, Y.; Chen, B.; Li, X.; Niu, Z.; Jiang, S.; Li, J.; Cui, J.; Yu, Y. Monitoring Aquatic Vegetation Distribution of Taihu Lake from Sentinel-2 and Random Forest Algorithm. Monit. Forewarning 2023, 15, 42–49. [Google Scholar]
  28. Gao, H.; Li, R.; Shen, Q.; Yao, Y.; Shao, Y.; Zhou, Y.; Li, W.; Li, J.; Zhang, Y.; Liu, M. Deep-Learning-Based Automatic Extraction of Aquatic Vegetation from Sentinel-2 Images—A Case Study of Lake Honghu. Remote Sens. 2024, 16, 867. [Google Scholar] [CrossRef]
  29. Xin, Y.; Luo, J.; Xu, Y.; Sun, Z.; Qi, T.; Shen, M.; Qiu, Y.; Xiao, Q.; Huang, L.; Zhao, J.; et al. SSAVI-GMM: An Automatic Algorithm for Mapping Submerged Aquatic Vegetation in Shallow Lakes Using Sentinel-1 SAR and Sentinel-2 MSI Data. IEEE Trans. Geosci. Remote Sens. 2024, 62, 4416610. [Google Scholar] [CrossRef]
  30. Xin, Y.; Luo, J.; Zhai, J.; Wang, K.; Xu, Y.; Qin, H.; Chen, C.; You, B.; Cao, Q. An Automatic Algorithm for Mapping Algal Blooms and Aquatic Vegetation Using Sentinel-1 SAR and Sentinel-2 MSI Data. Land 2025, 14, 592. [Google Scholar] [CrossRef]
  31. Nanjing Institute of Geography and Limnology, Chinese Academy of Sciences. China Lake Survey Report; Science Press: Beijing, China, 2019. [Google Scholar]
  32. The Editorial Committee. China National Geographic Atlas; Encyclopedia of China Publishing House: Beijing, China, 2010. [Google Scholar]
  33. Zhang, Z.; Zhang, M.; Xiao, W.; Wang, W.; Xiao, Q.; Wang, Y.; Li, X. Analysis of Temporal and Spatial Variations in NDVI of Aquatic Vegetation in Lake Taihu. J. Remote Sens. 2018, 22, 324. [Google Scholar] [CrossRef]
  34. Wang, S.; Gao, Y.; Li, Q.; Gao, J.; Zhai, S.; Zhou, Y.; Cheng, Y. Long-Term and Inter-Monthly Dynamics of Aquatic Vegetation and Its Relation with Environmental Factors in Taihu Lake, China. Sci. Total Environ. 2019, 651, 367–380. [Google Scholar] [CrossRef]
  35. Liu, L.; Xiong, J.; Zhang, Y.; Lu, Y.; Cai, X. Trajectory of Aquatic Vegetation Cover in Honghu Lake in Recent 23 Years Based on Multi-temporal Image Classification. Resour. Environ. Yangtze Basin 2025, 34, 126–139. [Google Scholar]
  36. Editorial Committee for “Physical Geography of China”, Chinese Academy of Sciences. Physical Geography of China: Climate; Science Press: Beijing, China, 1984. [Google Scholar]
  37. Ding, Y.; Wang, S.; Zheng, J.; Wang, H.; Yang, X. The Climate of China; Physical Geography of China, Science Press: Beijing, China, 2012. [Google Scholar]
  38. Taihu Basin Authority of Ministry of Water Resources. 2024 Water Resources Bulletin for the Taihu Basin and Rivers in Southeast China. 2025. Available online: https://www.tba.gov.cn/slbthlyglj/upload/63f5e07f-4a28-462f-bc9c-fbb41aceaf21.pdf (accessed on 30 September 2025).
  39. Zhao, K.; Zhou, Y.; Jiang, Z.; Hu, J.; Zhang, X.; Zhou, J.; Wang, G. Changes of Aquatic Vegetation in Lake Taihu Since 1960s. J. Lake Sci. 2017, 29, 351–362. [Google Scholar] [CrossRef]
  40. Tang, J.; Tian, G.; Wang, X.; Wang, X.; Song, Q. The Methods of Water Spectra Measurement and Analysis I:Above-Water Method. J. Remote Sens. 2004, 8, 37–44. [Google Scholar] [CrossRef]
  41. Long, T.; Jiao, W.; He, G. RPC Estimation via -Norm-Regularized Least Squares (L1LS)1. IEEE Trans. Geosci. Remote Sens. 2015, 53, 4554–4567. [Google Scholar] [CrossRef]
  42. Long, T.; Jiao, W.; He, G.; Zhang, Z. A Fast and Reliable Matching Method for Automated Georeferencing of Remotely-Sensed Imagery. Remote Sens. 2016, 8, 56. [Google Scholar] [CrossRef]
  43. Long, T.; Jiao, W.; He, G.; Yin, R.; Wang, G.; Zhang, Z. Block Adjustment with Relaxed Constraints from Reference Images of Coarse Resolution. IEEE Trans. Geosci. Remote Sens. 2020, 58, 7815–7828. [Google Scholar] [CrossRef]
  44. Long, T.; Jiao, W.; He, G.; Wang, G.; Zhang, Z. Digital Orthophoto Map Products and Automated Generation Algorithms of Chinese Optical Satellites. J. Remote Sens. 2023, 27, 635–650. [Google Scholar] [CrossRef]
  45. Vermote, E.; Tanre, D.; Deuze, J.; Herman, M.; Morcette, J.J. Second Simulation of the Satellite Signal in the Solar Spectrum, 6S: An Overview. IEEE Trans. Geosci. Remote Sens. 1997, 35, 675–686. [Google Scholar] [CrossRef]
  46. Kotchenova, S.Y.; Vermote, E.F.; Matarrese, R.; Frank, J.; Klemm, J. Validation of a Vector Version of the 6S Radiative Transfer Code for Atmospheric Correction of Satellite Data. Part I: Path Radiance. Appl. Opt. 2006, 45, 6762–6774. [Google Scholar] [CrossRef]
  47. Kotchenova, S.Y.; Vermote, E.F. Validation of a Vector Version of the 6S Radiative Transfer Code for Atmospheric Correction of Satellite Data. Part II. Homogeneous Lambertian and Anisotropic Surfaces. Appl. Opt. 2007, 46, 4455–4464. [Google Scholar] [CrossRef]
  48. Kotchenova, S.Y.; Vermote, E.F.; Levy, R.; Lyapustin, A. Radiative Transfer Codes for Atmospheric Correction and Aerosol Retrieval: Intercomparison Study. Appl. Opt. 2008, 47, 2215–2226. [Google Scholar] [CrossRef] [PubMed]
  49. Berk, A.; Acharya, P.K.; Bernstein, L.S.; Anderson, G.P.; Lewis, P.; Chetwynd, J.H.; Hoke, M.L. Band Model Method for Modeling Atmospheric Propagation at Arbitrarily Fine Spectral Resolution. U.S. Patent US7433806B2, 7 October 2008. [Google Scholar]
  50. Berk, A.; Conforti, P.; Kennett, R.; Perkins, T.; Hawes, F.; van den Bosch, J. MODTRAN6: A Major Upgrade of the MODTRAN Radiative Transfer Code. In Proceedings of the Algorithms and Technologies for Multispectral, Hyperspectral, and Ultraspectral Imagery XX, Baltimore, MD, USA, 5–9 May 2014; Volume 9088, pp. 113–119. [Google Scholar] [CrossRef]
  51. Li, W.; Huang, Y.; Shen, Q.; Yao, Y.; Xu, W.; Shi, J.; Zhou, Y.; Li, J.; Zhang, Y.; Gao, H. Assessment of Seven Atmospheric Correction Processors for the Sentinel-2 Multi-Spectral Imager over Lakes in Qinghai Province. Remote Sens. 2023, 15, 5370. [Google Scholar] [CrossRef]
  52. Franz, B. Remote Sensing Reflectance and Derived Products: MODIS & VIIRS. Available online: https://modis.gsfc.nasa.gov/sci_team/meetings/201606/presentations/plenary/franz.pdf (accessed on 1 April 2025).
  53. Zhou, Z.; Rahman Siddiquee, M.M.; Tajbakhsh, N.; Liang, J. UNet++: A Nested U-Net Architecture for Medical Image Segmentation. In Proceedings of the Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support; Stoyanov, D., Taylor, Z., Carneiro, G., Syeda-Mahmood, T., Martel, A., Maier-Hein, L., Tavares, J.M.R., Bradley, A., Papa, J.P., Belagiannis, V., et al., Eds.; Springer: Cham, Switzerland, 2018; pp. 3–11. [Google Scholar] [CrossRef]
Figure 1. Overview of the study area: The Yangtze River is divided into three main segments—its upper reaches lie west of Yichang, the middle reaches extend from Yichang to Hukou, and the lower reaches stretch eastward from Hukou.
Figure 1. Overview of the study area: The Yangtze River is divided into three main segments—its upper reaches lie west of Yichang, the middle reaches extend from Yichang to Hukou, and the lower reaches stretch eastward from Hukou.
Remotesensing 17 03817 g001
Figure 2. Table of Chinese domestic satellite imagery for Lake Taihu during April–October 2024.
Figure 2. Table of Chinese domestic satellite imagery for Lake Taihu during April–October 2024.
Remotesensing 17 03817 g002
Figure 3. Spatial distribution of aquatic vegetation field validation sites.
Figure 3. Spatial distribution of aquatic vegetation field validation sites.
Remotesensing 17 03817 g003
Figure 4. Reflectance spectral characteristics of various surface types.
Figure 4. Reflectance spectral characteristics of various surface types.
Remotesensing 17 03817 g004
Figure 5. Evaluation of atmospheric correction.
Figure 5. Evaluation of atmospheric correction.
Remotesensing 17 03817 g005
Figure 6. Spectral response functions (SRFs) of various satellite sensors. The semi-transparent curves represent the SRFs of individual cameras within sensors equipped with multiple cameras. The opaque curves denote the averaged SRFs derived from all cameras of each corresponding sensor.
Figure 6. Spectral response functions (SRFs) of various satellite sensors. The semi-transparent curves represent the SRFs of individual cameras within sensors equipped with multiple cameras. The opaque curves denote the averaged SRFs derived from all cameras of each corresponding sensor.
Remotesensing 17 03817 g006
Figure 7. Sensor consistency evaluation based on spectral response function: from left to right are the blue, green, red, and near-infrared bands.
Figure 7. Sensor consistency evaluation based on spectral response function: from left to right are the blue, green, red, and near-infrared bands.
Remotesensing 17 03817 g007
Figure 8. Sensor consistency evaluation based on imagery: from left to right are the blue, green, red, and near-infrared bands.
Figure 8. Sensor consistency evaluation based on imagery: from left to right are the blue, green, red, and near-infrared bands.
Remotesensing 17 03817 g008
Figure 9. Model learning curves.
Figure 9. Model learning curves.
Remotesensing 17 03817 g009
Figure 10. Model evaluation based on visually interpreted aquatic vegetation classification results.
Figure 10. Model evaluation based on visually interpreted aquatic vegetation classification results.
Remotesensing 17 03817 g010
Figure 11. Model evaluation based on field validation of aquatic vegetation. Circles denote buffer zones around field validation sites. The background class is shown with an inverted color scheme, and aquatic vegetation classes are displayed in darker tones for clearer contrast.
Figure 11. Model evaluation based on field validation of aquatic vegetation. Circles denote buffer zones around field validation sites. The background class is shown with an inverted color scheme, and aquatic vegetation classes are displayed in darker tones for clearer contrast.
Remotesensing 17 03817 g011
Figure 12. Aquatic vegetation classification results of different methods.
Figure 12. Aquatic vegetation classification results of different methods.
Remotesensing 17 03817 g012
Figure 13. Spatiotemporal variations of aquatic vegetation in Lake Taihu from 2020 to 2025.
Figure 13. Spatiotemporal variations of aquatic vegetation in Lake Taihu from 2020 to 2025.
Remotesensing 17 03817 g013
Figure 14. Monthly variations of aquatic vegetation in Lake Taihu during 2024 and 2025.
Figure 14. Monthly variations of aquatic vegetation in Lake Taihu during 2024 and 2025.
Remotesensing 17 03817 g014
Table 1. Life-form composition of aquatic vegetation across different lakes.
Table 1. Life-form composition of aquatic vegetation across different lakes.
LakeSubmerged Vegetation (SV)Floating-Leaved (FV)Emergent Vegetation (EV)
Lake Taihu
Lake Chaohu×
Lake Honghu×
Lake Dianchi×
Lake Dianshanhu
Lake Caohai×
Table 2. Spectral band configuration of Chinese domestic satellites.
Table 2. Spectral band configuration of Chinese domestic satellites.
SensorGF-1 WFVGF-6 WFVHJ-2A CCDHJ-2B CCD
Band
B1450–520450–520450–520450–520
B2520–590520–590520–590520–590
B3630–690630–690630–690630–690
B4770–890770–890770–890770–890
B5/690–730690–730690–730
B6/730–770//
B7/400–450//
B8/590–630//
Note: “/” denotes the absence of the corresponding spectral band in that sensor.
Table 3. Remote sensing interpretation characteristics of different surface types.
Table 3. Remote sensing interpretation characteristics of different surface types.
ClassTrue Color CompositeFalse Color CompositeDescription
Submerged Vegetation (SV)Remotesensing 17 03817 i001Remotesensing 17 03817 i002Appears dark green or nearly black in true color composites, and dark red or black in false color composites, exhibiting a rough surface-like distribution.
Floating-leaved Vegetation (FV) Remotesensing 17 03817 i003Remotesensing 17 03817 i004Appears light green in true color composites, and pale pink in false color composites, distributed in rough surface-like or fragmented patterns along the lakeshore.
Emergent Vegetation (EV) Remotesensing 17 03817 i005Remotesensing 17 03817 i006Appears bright to dark green tones in true color composites, and bright pink to red tones in false color composites, distributed in rough surface-like patterns along the lakeshore.
WaterRemotesensing 17 03817 i007Remotesensing 17 03817 i008Appears blue or black in both true, and false color composites, showing a smooth surface-like distribution.
Algal Bloom (AB)Remotesensing 17 03817 i009Remotesensing 17 03817 i010Appears bright green in true color composites, and bright pink in false color composites, showing a smooth flocculent distribution.
Table 4. Selection of input parameters.
Table 4. Selection of input parameters.
Exp.Input BandInput Block SizemIoU ± std (%)mPA ± std (%)
Final ModelNIR + R + G 256 × 256 77.49 ± 1.15 84.65 ± 1.30
+ BlueNIR + R + G + B 256 × 256 76.06 ± 0.22 83.48 ± 0.42
+ NDVINIR + R + G + NDVI 256 × 256 72.73 ± 0.96 81.30 ± 0.51
+ Input SizeNIR + R + G 128 × 128 73.69 ± 1.49 81.42 ± 1.26
Note: bold text indicates input-parameter settings that differ from those used in the final model configuration.
Table 5. Selection of model components.
Table 5. Selection of model components.
Exp.EncoderDecoderLossmIoU ± std (%)mPA ± std (%)
Final ModelEfficientNet-B5U-Net++Dice + Focal77.49 ± 1.1584.65 ± 1.30
LossEfficientNet-B5U-Net++Cross Entropy76.51 ± 1.0883.67 ± 0.92
EncoderResNet-34U-Net++Dice + Focal76.90 ± 1.1684.36 ± 1.06
DecoderEfficientNet-B5U-NetDice + Focal76.41 ± 0.0783.24 ± 0.16
Note: bold text highlights model-component modifications relative to the final model architecture.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Shao, Y.; Shen, Q.; Yao, Y.; Wang, X.; Zhao, H.; Gao, H.; Zhou, Y.; Zhang, H.; Gong, Z. Deep Learning-Based Classification of Aquatic Vegetation Using GF-1/6 WFV and HJ-2 CCD Satellite Data. Remote Sens. 2025, 17, 3817. https://doi.org/10.3390/rs17233817

AMA Style

Shao Y, Shen Q, Yao Y, Wang X, Zhao H, Gao H, Zhou Y, Zhang H, Gong Z. Deep Learning-Based Classification of Aquatic Vegetation Using GF-1/6 WFV and HJ-2 CCD Satellite Data. Remote Sensing. 2025; 17(23):3817. https://doi.org/10.3390/rs17233817

Chicago/Turabian Style

Shao, Yifan, Qian Shen, Yue Yao, Xuelei Wang, Huan Zhao, Hangyu Gao, Yuting Zhou, Haobin Zhang, and Zhaoning Gong. 2025. "Deep Learning-Based Classification of Aquatic Vegetation Using GF-1/6 WFV and HJ-2 CCD Satellite Data" Remote Sensing 17, no. 23: 3817. https://doi.org/10.3390/rs17233817

APA Style

Shao, Y., Shen, Q., Yao, Y., Wang, X., Zhao, H., Gao, H., Zhou, Y., Zhang, H., & Gong, Z. (2025). Deep Learning-Based Classification of Aquatic Vegetation Using GF-1/6 WFV and HJ-2 CCD Satellite Data. Remote Sensing, 17(23), 3817. https://doi.org/10.3390/rs17233817

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop