Previous Article in Journal
Chloroplast Hibernation-Promoting Factor PSRP1 Prevents Ribosome Degradation Under Darkness Independently of 100S Dimer Formation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Self-Supervised Learning and Multi-Sensor Fusion for Alpine Wetland Vegetation Mapping: Bayinbuluke, China

1
State Key Laboratory of Ecological Safety and Sustainable Development in Arid Lands, Xinjiang Institute of Ecology and Geography, Chinese Academy of Sciences, Urumqi 830011, China
2
University of Chinese Academy of Sciences, Beijing 100049, China
3
China-Kazakhstan Joint Laboratory for RS Technology and Application, Al-Farabi Kazakh National University, Almaty 050012, Kazakhstan
4
CAS Research Center for Ecology and Environment of Central Asia, Urumqi 830011, China
5
National Engineering Technology Research Center for Desert-Oasis Ecological Construction, Urumqi 830011, China
6
Department of Electrical, Computer and Biomedical Engineering, University of Pavia, 27100 Pavia, Italy
*
Author to whom correspondence should be addressed.
Plants 2025, 14(20), 3153; https://doi.org/10.3390/plants14203153 (registering DOI)
Submission received: 1 September 2025 / Revised: 1 October 2025 / Accepted: 6 October 2025 / Published: 13 October 2025
(This article belongs to the Special Issue Computer Vision Techniques for Plant Phenomics Applications)

Abstract

Accurate mapping of wetland vegetation is essential for ecological monitoring and conservation, yet it remains challenging due to the spatial heterogeneity of wetlands, the scarcity of ground-truth data, and the spread of invasive species. Invasive plants alter native vegetation patterns, making their early detection critical for preserving ecosystem integrity. This study proposes a novel framework that integrates self-supervised learning (SSL), supervised segmentation, and multi-sensor data fusion to enhance vegetation classification in the Bayinbuluke Alpine Wetland, China. High-resolution satellite imagery from PlanetScope-3 and Jilin-1 was fused, and SSL methods—including BYOL, DINO, and MoCo v3—were employed to learn transferable feature representations without extensive labeled data. The results show that SSL methods exhibit consistent variations in classification performance, while multi-sensor fusion significantly improves the detection of rare and fragmented vegetation patches and enables the early identification of invasive species. Overall, the proposed SSL–fusion strategy reduces reliance on labor-intensive field data collection and provides a scalable, high-precision solution for wetland monitoring and invasive species management.

1. Introduction

Wetland ecosystems play a crucial role in maintaining ecological balance and biodiversity [1]. Therefore, accurate mapping of these landscapes is vital for conservation and sustainable development [2]. Remote sensing has been widely used for wetland vegetation monitoring and hyperspectral imagery (HSI) offers distinct advantages through its fine spectral resolution which enables broad vegetation discrimination and detection of stable ecological changes [3]. However, the effectiveness of HIS is often limited by insufficient ground-truth data, particularly in heterogeneous wetlands with different land cover types [4]. Additional challenges include high spatial heterogeneity, temporal variability, fluctuating water levels, and spectrally similar plant species [5].
Another major challenge in wetland mapping arises from invasive plant species (IPS), which displace native vegetation, alter hydrological processes, and reduce biodiversity [6,7]. Moreover, from a remote sensing perspective, invasive plants are difficult to identify because their spectral similarity to co-occurring vegetation can obscure class separability. Patchy distribution and rapid spread also introduce temporal variability that complicates ecological monitoring [8,9]. Early and accurate identification of IPS is essential for the precision of wetland vegetation maps. This helps direct focused conservation and restoration efforts [10,11].
Although hyperspectral imagery (HSI) enhances vegetation discrimination by capturing hundreds of narrow bands [12] and spectral–spatial models to achieve better classification accuracy [13], traditional supervised learning (SL) approaches such as Support Vector Machines (SVM) and Random Forests (RF) have been used to classify wetlands [14]. In wetlands [15], these challenges highlight the limitations of purely supervised approaches and motivate the development of alternative strategies [16].
A viable approach is self-supervised learning (SSL), which mitigates data scarcity by deriving features from unlabeled data [17] and utilizing pretext activities to extract features pertinent for further classification [18]. Recent advancements in contrastive and predictive self-supervised learning, including global–local alignment and multiscale context modeling, have demonstrated robust efficacy in highly changeable land-cover environments. Methods such as BYOL [19], DINOv2 [20], and MoCo v3 [21] have exhibited robust performance in remote sensing applications characterized by a scarcity of labeled data [22]. Applying SSL to hyperspectral wetland imagery enables the extraction of high-level feature representations that capture both spatial and spectral vegetation patterns, along with other environmental features [23]. These representations are generated directly from the data, eliminating the need for predefined labels—particularly advantageous in wetland mapping, where annotated datasets are often unavailable [24]. These SSL-based features can then be integrated into supervised learning pipelines, such as DeepLabV3, a widely used semantic segmentation framework in remote sensing [25]. DeepLabV3 employs an atrous convolution architecture that facilitates the acquisition of multiscale contextual and spatial details, which is particularly advantageous due to the limited availability of annotated wetland datasets [26].
Additionally, SSL, SL, and data fusion (DF) are an essential process for improving classification accuracy. Such complementary spectral, spatial, and temporal resolutions can be separately accessed in different remote sensing datasets, allowing a more comprehensive representation of wetland ecosystems [27]. High-resolution imagery captures fine structural details [28], whereas high-temporal-resolution data elucidate seasonal and interannual changes [29]. Multispectral data add unique spectral information that improves the separability of vegetation types that may otherwise appear similar in a single dataset [30]. By integrating these heterogeneous sources, data fusion (DF) produces more robust and precise models for wetland vegetation classification [31]. For example, fusion hyperspectral dataset can enhance both spectral fidelity and spatial detail [32], particularly when supported by transformer-based spatial–spectral networks [33].
However, data-fused imaging frequently suffers from noise, particularly in wetlands, where mixed pixels and atmospheric influences compromise signal integrity. This work employs EN2N, which is a self-supervised Noise2Noise technique tailored for hyperspectral imaging. Analogous to SPEND, which eliminates spatially correlated noise in the absence of clear references [34], EN2N directly denoises noisy observations, maintaining spectral–spatial detail, and markedly enhancing the robustness and precision of wetland vegetation categorization.
This study demonstrated a framework that combines self-supervised learning (SSL) to extract supervised learning (SL) features to classify and utilize data fusion (DF) techniques to create better maps of wetland vegetation. There are some advantages to such a combination: with the use of SSL, it is possible to have fewer dependencies on manually annotated data and SL has a good classification capability, while DF improves the performance of the model due to synthesized complementary information by multiple sensors. Together, these methods offer a more reliable and accurate approach to monitoring and protecting wetlands. In addition, the improvement in methodology improves the framework for long-term monitoring, sustainable wetland management, and biodiversity conservation. These tools are badly required in light of the increasing human pressures from climate, land use, and human development. Furthermore, the research methodology also provides a viable solution to current problems and offers opportunities for future studies in remote sensing and ecological surveillance.

2. Results

2.1. Implementation Details

We employ a ResNet-50 as the online encoder for self-supervised pretraining. The encoder begins with a 7 × 7 convolution (stride =2 × 2, padding =3 × 3) whose input channels are adapted to the dataset’s full spectral depth . Given an input patch X   ϵ   x 64 x 64 , the stem applies weights W 0   64 x x 7 x 7 , after which features traverse the ResNet bottleneck blocks and an adaptive average pooling layer to yield a 2048-D representation, as shown in Equation (1).
z = f 0 x ϵ   2048
A projection head g . is used only during SSL (BYOL, DINO, MoCo v3), producing p = g z . A momentum (key) encoder mirrors the online encoder and is updated by exponential moving average (EMA) with momentum m = 0.99 , which is elaborated in Equation (2):
θ t + 1 = m θ t + 1 m θ t
In Equation (2) θ t denotes the online network parameters at iteration t, while θ t represents the parameters of the target network, updated using an exponential moving average (EMA):
On the other hand, SSL training minimizes a contrastive/predictive objective over two stochastic views T1 and T2 of the same patch, which is represented in Equation (3),
L S S L = Ε Χ 1 , Τ 1 , Τ 2 l g ϕ f 0 Τ 1 x , s g g ϕ f 0 Τ 2 x
where l g f 0 T 1 x , s g g f 0 T 2 x is the method-specific loss (e.g., cosine/predictive for BYOL, InfoNCE for MoCo v3, centering/sharpening for DINO) and s g g f 0 T 2 x denotes stop-gradient as appropriate. We split the dataset into 80%training and 20%validation, retaining the natural per-class distribution (see Table 2). After SSL, encoder weights θ are transferred downstream.
For semantic segmentation, the SSL-pretrained backbone initializes DeepLabV3, which is fine-tuned on the same 64 × 64 labeled patches. Let H × W be the spatial size and c l s the number of classes with logits u i , c , and one hot labels y i , c at pixel Plants 14 03153 i001. We minimize the loss in Equation (4):
L s e g = i = 1 H   ×   W c = 1 C c l s y i , c l o g s o f t m a x u i , c
Final predictions on the Alpine Wetland imagery are generated using a sliding window over the full-resolution mosaics to ensure complete coverage.
To make the distinction explicit, we define a task factor represented in Equation (5):
f x = z = f 0                                                                                         S S L   P r e t r a i n i n g r e p r e s e n t a t i o n   l e a r n i n g y = a r g m a x   \ s o f t m a x ( h 0 f 0 x ) c       D e e p l a b V 3 : P i x e l w i s e   l a b e l l i n g
where h is the segmentation head and all other hyperparameters follow the defaults of the respective SSL methods.

2.2. Comparing with Other SSL Methods

On the Jilin-1 dataset, MoCo v3 achieved the highest accuracy of 86%TOP 1 OA, followed by DINO at 83% TOP 1 OA and BYOL at 80% TOP 1 OA. On the PlanetScope-3 dataset, MoCo v3 reported the best result with 97%TOP 1 OA, while BYOL and DINO reached 83%TOP 1 OA and 77% TOP 1 OA, respectively. For the fused dataset, BYOL achieved 95% TOP 1 OA, slightly lower than MoCo v3 at 96%, while DINO obtained 87%, as shown in Figure 1.
These results indicate that performance varied across methods and datasets. MoCo v3 consistently produced the highest scores on the single-sensor datasets (Jilin-1 and PlanetScope-3), while BYOL achieved the top result on the fused dataset.

2.3. Fine Tune SSL by DeeplabV3

To achieve high-precision wetland vegetation mapping, we developed a fine-tuning pipeline built on DeepLabV3+ with a ResNet-50 backbone which is graphically shown in Figure 2. The workflow begins with multi-sensor satellite imagery, which is preprocessed and segmented into patches before being used for self-supervised pretraining with BYOL, DINO, or MoCo v3. These pretrained encoders provide robust feature representations, which are then transferred into DeepLabV3+ for supervised fine-tuning using annotated vegetation masks. The final output is a detailed segmentation map, where vegetation classes are clearly delineated and non-informative background pixels (class 0) are excluded from metrics. This approach demonstrates the benefits of SSL in boosting downstream segmentation performance in complex wetland environments.
Moreover, the refined DeepLabV3 classification outcomes were obtained using different SSL approaches and datasets from the Bayinbuluke Alpine Wetland. The original Jilin-1 satellite, PlanetScope-3, and data fusion image are illustrated as follows: (a) Jilin-1 (BYOL), (b) Planet Scope 3 (BYOL), or (c) data fusion (BYOL); (d) Jilin-1 (MOCOv3), (e) Planet Scope 3 (MOCOv3), or (f) data fusion (MOCOv3); (g) Jilin-1 DINO, (h) Planet Scope 3 (DINO), or (i) data fusion (DINO). In the same way, the segmentation maps of MoCo v3 over Jilin-1, PlanetScope-3, and the fused dataset are provided in panel (g) to (i) in Figure 3.
Overall, the fused source delivers highly textured spatial information and enhances discrimination between vegetation classes—especially when distinguishing grassland associations (Elymus dahuricus and Pedicularis spp.) and differentiating meadow types. Quantitatively, BYOL provides the strongest segmentation quality overall, achieving the highest overall accuracy (O/A) on all three datasets and the best mIoU on Jilin-1 and the fused set, indicating sharper fine-scale boundaries. On PlanetScope-3, DINO attains the highest mIoU (65%), consistent with smoother, well-delineated boundaries on that sensor, although its O/A is lower than BYOL. MoCo v3 is competitive but slightly behind across sensors; its lower mIoU—particularly on PlanetScope-3—suggests under-segmentation of smaller classes when DeepLabV3 is trained on single-sensor inputs and accuracies metrices. The comparison indicates that DF improves class separability for most methods, with the BYOL-initialized model preserving the best balance between accuracy and boundary quality.
On the Jilin-1 dataset, BYOL achieved the highest performance with 93% overall accuracy and 64% mIoU, while DINO reached 76% overall accuracy and 56% mIoU, and MoCo v3 obtained 72% overall accuracy and 58% mIoU. For PlanetScope-3, BYOL again obtained the best overall accuracy (92%), whereas DINO produced the highest mIoU (65%); MoCo v3 followed with 86% overall accuracy and 61% mIoU. On the fused dataset, BYOL delivered the top scorers with 88% overall accuracy and 64% mIoU, closely followed by MoCo v3 at 87% and 63%, while DINO achieved the same overall accuracy (87%) but a lower mIoU of 57%, as shown in Table 1.
However, radar chart (Figure 4) represents the overall accuracy between BYOL, DINO, and MoCo v3 on Jilin-1, PlanetScope-3, and fused datasets. The visualization allows us to see the performance differences according to various datasets and methods. The fused dataset gave the best result on the BYOL model (95%), and MoCo v3 yielded the best results on the single-sensor datasets (86% on Jilin-1 and 97% on PlanetScope-3). This indicates that SSL performance is dataset-specific, and MoCo v3 is better modeled on single-source datasets, while BYOL is good at multiple-sensor fusion.

3. Discussion

The results demonstrate that the performance of SSL methods is strongly influenced by both the input dataset and the downstream task. MoCo v3 consistently achieved the highest accuracies on the single-sensor datasets, particularly with PlanetScope-3 imagery, where its contrastive learning framework benefited from high temporal frequency and moderate spatial resolution. This finding is consistent with earlier research showing that contrastive SSL systems are especially effective when the spectral dimensionality of medium-resolution satellite images is limited but temporal variability is large [19,35]. This indicates that negative sample-based approaches are well suited for scenarios with limited spectral dimensionality but strong temporal variability. In contrastive SSL, a negative sample-based method is used whereby the model is trained to bring augmented views of the same image (positives) closer together in the feature space while pushing apart the representations of different images (negatives). This mechanism prompts the encoder to learn highly discriminative features that generalize well across diverse temporal measurements. By contrast, non-contrastive SSL designs such as BYOL rely on asymmetric network architectures and momentum updates rather than explicit negative pairs. The strong results of MoCo v3 on PlanetScope-3 therefore underline the advantage of negative sample-based contrastive learning in cases where spectral richness is low but temporal dynamics are high.
In contrast, BYOL had better performance on the fused dataset, demonstrating that predictive self-supervised learning without negative pairs works best in the high-dimension spectral–spatial setting. Similar trends have been observed in hyperspectral and multisensor applications, where redundancy among channels reduces the benefits of contrastive learning [17]. In contrast to methods that risk collapsing features when features are redundant, BYOL is actually able to make good use of complementary spectral and spatial information, without any requirement to define a negative set. This property favors BYOL being precisely appropriate in data fusion applications in the context of both redundancy and complementary information among sensors.
DINO exhibited mixed performance. While it underperformed on Jilin-1 and showed instability across datasets, it achieved the highest mIoU on PlanetScope-3, suggesting that self-distillation can help preserve boundary quality under certain imaging conditions. Similar improvements in boundary delineation have been reported in remote sensing segmentation tasks [36,37]. Although the overall sensitivity to dataset characteristics reduces its robustness compared with BYOL and MoCo v3.
The segmentation results further highlight the benefits of SSL pretraining. Across all methods, SSL-initialized DeepLabV3 models produced sharper boundaries and clearer class separation than purely supervised training with limited ground-truth, consistent with findings in ecological and agricultural remote sensing [38,39]. Data fusion consistently improved the identification of minority and fragmented classes, such as Elymus- and Pedicularis-dominated patches, wet meadows, and invasive species. These improvements are ecologically significant, as the ability to delineate rare vegetation types and invasive species is critical for biodiversity monitoring and conservation management. Prior research has pointed out the difficulty of targeting invasive species because they share spectral characteristics with other vegetation present [8,10,40]. The fact that fusion-SSL methods did work in this case illustrates the fact that it is possible to overcome this difficulty.
This research has several limitations although the results are promising. The use of only one point in time (June 2024) represents the vegetation during its peak season without accounting for the dynamics of phenology throughout the year. These limitations could prevent the identification of species that have a characteristic seasonal pattern, such as invasive plants that grow at varying stages. It could also be more relevant to include multi-temporal datasets that have been used to perform good vegetation classification and invasion species detection [41,42], to bring a broader perspective of the wetland dynamics. The ground-truth volume of data is also a limitation as it is relatively small. Although through the use of SSL, there is less reliance on annotated samples, the small number of field observations may influence the overall validity of the framework in its application to other wetlands.
These difficulties should be addressed in future studies by increasing field sampling and using semi-supervised or active learning to tradeoff between the cost of annotation and the quality of validation. To further enhance robustness of the sensors in heterogeneous wetlands, integration of complementary sensors including synthetic aperture radar (SAR) and LiDAR, which record structural and backscatter data, can be employed. The transferability of the framework across various ecological and imaging conditions would also be possible by applying the framework to different wetland systems. The application of transfer learning, UAV validation, and closer integration with biodiversity monitoring programs would scale the approach and make it more applicable for long-term conservation and ecological management.

4. Material and Methods

4.1. Study Area

The study area is the Bayinbuluke Wetland, located in the Kaidu River Basin in the central Tianshan Mountains of Xinjiang Uyghur Autonomous Region, northwest China, as shown in a Figure 5. The Bayinbuluke Wetland (b) subregion used for classification and analysis is identified in Figure 5c as the focused area of Bayinbuluke. Moreover, it extends between 42.78°to 43.23° latitude and 83.68° to 84.27° longitude with an average elevation of 2400 m, characteristics of a high-altitude alpine basin. The region has a cold temperate continental climate, which is manifested by long, cold winters and short, warm summers. The average monthly temperatures are as low as −20 °C in January to as high as 15 °C in July, and the mean annual precipitation is 270 mm, mostly from May to September. Potential evapotranspiration exceeds more than 1000 mm annually, and glacial meltwater together with seasonal rainfall, sustains wetlands inundation and soil saturation, peaking in spring and early summer.
Bayinbuluke Wetland is one of the largest alpine wetlands in China and forms part of the Bayanbulak Grassland National Nature Reserve. It supports a mosaic of ecosystems including alpine meadows, marshes and rivers, and shallow lakes that provide critical breeding habitats for migrating birds, notably the endangered Black-necked Crane (Grus nigricollis). Vegetation includes cold-adapted grasses, herbaceous plants, and lowland meadows tolerant of flooding. The dominant land cover consists of alpine grasslands (~65%), marsh wetlands (~20%), and open water (~5%), along with limited rocky and gravel surfaces.
For this study, a subregion of Bayinbuluke Wetland (Figure 5c was selected for land cover classification using remote sensing and machine learning. Twelve vegetation and land cover classes were defined, including natural meadows, open water, developed surfaces, and herbaceous vegetation types, some of which represent known invasive species. This scheme enables accurate delineation of habitat composition and structure, forming the basis for subsequent segmentation modeling and ecological analysis.

4.2. Data and Processing

4.2.1. Satellite Imagery

We used two high-resolution satellite datasets PlanetScope-3 (3 m) and Jilin-1 (4 m), as the main inputs for wetland classification. The PlanetScope-3 constellation captures imagery across eight spectral bands, including coastal blues, blue, green, yellow, red, red edge, and two near-infrared (NIR), while the Jilin-1 satellite provides four bands including blue, green, red, and near-infrared (NIR). The overall data was obtained in June 2024, which was the peak of the growing season in Bayinbuluke when vegetation differences are the most significant. The data on satellite imagery, topography, proximity, and reference data employed in this study are reported in detail in Table 2.
Moreover, DF was performed to enhance the classification accuracy as well as the spectral richness by combining the two satellite sources to provide one dataset having 11 spectral bands to better classify the spectral information and greater spectral detail. Geometric correction, co-registration, and resampling to a common spatial resolution were performed to preprocess all imagery prior to fusion. This kind of combination of complementary spectral signatures enabled a more precise sorting of wetland vegetation and land cover categories. Moreover, overview of datasets and supporting information were used for wetland vegetation classification, including satellite imagery, topography, proximity, and reference data.
The data fusion output was then denoised through a convenient self-supervised SAR denoising model using the EN2N model, a self-supervised SAR denoising framework, which aims to represent a convolutional neural network (CNN) adapted during the current study to the domain of optical images preprocessing [43]. The denoising procedure maintained the spatial integrity of an image by minimizing the noise component, thus enhancing the quality of the dataset to which the classification would be applied, as described in Figure 6.

4.2.2. Topographic Information

Bayinbuluke Wetland in Xinjiang, located in the central Tianshan mountain of China, is one of the largest wetlands of the alpine regions in the country and is a component of the Bayanbulak Grassland National Nature Reserve. It lies at an average of about 2400 m above sea level with alpine meadows, marshes, shallow lakes, and riverine systems which offer important breeding areas to migratory birds such as the endangered Black-necked Crane (Grus nigricollis). The most common land cover types are alpine grasslands (more than 65 percent), marsh wetlands (more than 20 percent), open water (more than 5 percent), and small rocky and gravelly surfaces. The vegetation community comprises cold-adapted grasses, herbaceous plants, and meadow types partially adapted to low temperatures and seasonal flooding.
The basin topography is characterized by flat to gently undulating alluvial plains surrounded by steep mountains exceeding 4000 m. These slopes generate glacial and snowmelt runoff, which feeds into the Kaidu River and its tributaries, maintaining soil saturation and extensive marsh formation. Water retention is especially pronounced in the low-gradient central wetland zones, where slopes are generally <3°. Over centuries, meltwater deposition from Tianshan streams has shaped the meadows, marshes, and shallow river systems observed today.
The measurements of topographic variables, e.g., elevation, slope, and hydrological connectivity, were determined based on the digital elevation models (DEMs), i.e., Shuttle Radar Topography Mission (SRTM) and Advanced Spaceborne Thermal Emission and Reflection Radiometer Global DEM (ASTER GDEM). These datasets were used to classify the sub-regional land cover and for further ecological analysis.

4.2.3. Proximity Information

The Bayinbuluke Wetland is located along the headwaters of the Kaidu River, which is one of the major tributaries draining glacial meltwaters from surrounding highlands. Proximity to the Kaidu River and its distributaries is critical for seasonal hydrological processes in the wetland, particularly in periods of spring and early snowmelt of summer. The region is characterized by its remoteness and limited human disturbance, with few settlements scattered along the outer edge of the reserve. The area is lowly inhabited, remote, and not heavily disturbed by man, with occasional settlements found on the perimeter of the reserve. It is near Hejing County, about 200 km to the northwest, and about 500 km to Urumqi, the capital of Xinjiang. The ecological integrity of this area is maintained by the presence of the wetland with proximity to snow-capped mountains, and its distance from areas of intensive agriculture or urban expansion.
In this research, the proximity analysis was carried out with the help of the GPS coordinates gathered during field work in June 2024, coupled with the geographical characteristics of the area being taken from OpenStreetMap (OSM) and high-resolution imagery. These sources provide reliable spatial points of reference to map the boundaries of the wetlands and the surrounding streams, roads, and features of the landscape. Major hydrological references were made based on the Kaidu River that happens to wind through the basin. All the proximity layers were processed in GIS environment for analysis. Calculation of field GPS data and OSM data enabled the fine definition of wetland extents, buffer areas, and spatial relationships between vegetation zone and drivers of the environment, i.e., water bodies and elevation gradients.

4.2.4. Reference Data

This study used a combination of field-collected Global Positioning System (GPS) coordinates and vector data obtained via OpenStreetMap (OSM) as the reference data. The GPS points (the number of which was obtained across representative land cover types in June 2024) were collected by field teams and then cross-validated with high-resolution imagery and OSM-derived features in order to ensure spatial accuracy and contextual consistency in our data. Using this geospatial information, polygon shapefiles have been created manually by digitizing all 13 land cover and vegetation classes in the study area, including natural and artificial surfaces, as shown in Table 3.
These class polygons serve as the foundation for supervised learning utilizing the labeled regions, from which we extracted 64 × 64-pixel image patches from the fused 11-band data (PlanetScope-3 and Jilin-1) images so that each patch occupies the same spatial location as the label that it corresponds to. The procedure facilitated the creation of an organized dataset intended for subsequent picture classification and semantic segmentation via deep learning, while maintaining label integrity and minimizing overlap among classes at the boundaries between patches.
In order to take advantage of both sensors, the two datasets were co-registered and fused with each other spatially. The bands of PlanetScope-3 and Jilin-1 were merged together, and they formed a fused image with 11 distinctive spectral bands. The spectral richness and spatial details of the fused dataset help attain a high accuracy in classification of wetland vegetation.

4.2.5. Methodology Workflow

This research workflow includes three stages that are connected with one another, namely: data preparation, feature engineering, and modeling. The first step involved obtaining multi-source satellite imagery from PlanetScope-3 and Jilin-1, as well as ground-truth data in the form of shapefiles, GPS surveys, and raw field observations. The two satellite datasets were merged into an 11-band fused image by using a DF process, followed by preprocessing, co-registration, and cropping of the fused image along with their respective labels into 64 × 64 patches.
To train the dataset, 80 percent of the data were used as a training sample and 20 percent were used as a validation sample to ensure that a robust model is developed. The self-supervised learning (SSL) models BYOL, DINO, and MoCo v3 were trained on fused datasets and single-source datasets over 100 epochs with a ResNet-50 backbone during the feature engineering stage. Their representations were then gauged using a linear probe that had been trained in 100 epochs, and Top-k accuracies were collected to test the quality of features represented. The general workflow is shown in Figure 7.
During the last modeling phase, the trained SSL backbones were transferred to a DeepLabV3 segmentation model. The model was adjusted with 100 epochs on the labeled 64 × 64 patches, the ignore-0 index was used during the training to deal with the background pixels. The final resultant segmentation maps of the Bayinbuluke Wetland using the DeepLabV3-pretrained with the SSS were high-resolution. The overall accuracy (O/A) and mean Intersection over Union (mIoU) were used as metrics to determine model performance when comparing various methods of applying SSL and data fusion.

4.2.6. Computational Environment

All experiments were conducted on a workstation equipped with an Intel processor-based workstation with 64 GB RAM and NVIDIA GeForce RTX 4060, 8 GB VRAM, and Ubuntu 22.04 LTS. We conducted all experiments in PyCharm (2024.1), with the environment managed in Anaconda using Python 3.12.11 (conda-forge build, MSC v.1943, 64-bit),PyTorch 2.4.1 + cu124 was used to develop and evaluate the models with CUDA 12.4 turned on. The graphics card was identified as NVIDIA GeForce RTX 4060. The additional libraries were Torchvision 0.19.1+cu124, NumPy 1.26.4, Matplotlib 3.10.6, Lightly 1.5.20, and scikit-learn 1.7.2, Rasterio 1.4.3, and ArcGIS Pro were also used to complete geospatial data processing and visualization as summarized in Table 4.
In self-supervised pretraining (BYOL, DINO, MoCo v3), models were trained for 100 epochs with a batch size of 64, a learning rate of 1 × 10−4, and a weight decay of 1 × 10−4 were used. These conditions were typical of use in the literature and had been adjusted by preliminary experiments in order to stabilize training. Average pretraining time was approximately 1 h 45 min (±5 min) and linear evaluation took about 45 min (±5 min) per model.
For downstream semantic segmentation, the SSL-pretrained backbones were transferred to a DeepLabV3 model trained for 100 epochs, using the ignore-0 index to exclude background pixels. Fine-tuning required around 2 h (±10 min) per model. The conditioning of 64 × 64 patches was nearly real time (less than 1 s), enabling large-scale mapping of the Bayinbuluke Wetland.

5. Conclusions

This study demonstrated that integrating SSL with supervised segmentation and data fusion can deliver high-resolution wetland vegetation mapping with a significant improvement in accuracy. The tested techniques performed best on single-sensor and fused data, respectively, with MoCo v3 performing best on single-sensor data, whereas BYOL performed best on fused data, underscoring the need to tailor SSL approaches to data properties. DINO displayed promise with regard to establishing boundaries but lacked robustness across the datasets. The pretraining consistently improved the quality of segmentation, especially in minority and fragmented classes, whereas data fusion improved recognition of rare and invasive vegetation types. The findings mean that SSL limits the need to rely on large amounts of field labeled data and offers a feasible platform for complicated wetland environments. However, they have limitations regarding time coverage as well as the lack of ground-truth. The combination of SSL with the multi-sensor fusion has great potential as an avenue toward generating meaningful and operationally useful information on wetland vegetation.

Author Contributions

M.M.Z.: Conceptualization, investigation, visualization, writing—original draft, and writing—review and editing; A.S.: conceptualization, investigation, writing—review and editing, supervision, funding acquisition, and project administration; J.A.: writing—review and editing; E.Z.: Data curation; A.A.: writing—review and editing; W.L.: Data curation. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Tianshan Talent Development Program, grant number 2022TSYCCX0006, the National Natural Science Foundation of China, grant number 42371389, and the Western Young Scholars Project of the Chinese Academy of Sciences, grant number 2022-XBQNXZ-001.

Data Availability Statement

Data are available from the corresponding author upon reasonable request.

Acknowledgments

The authors gratefully acknowledge the financial support provided by the ANSO Scholarship for Young Talents and the University of Chinese Academy of Sciences (UCAS), their support made this research possible.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Chakraborty, S.K.; Sanyal, P.; Ray, R. Wetlands Ecology: Eco-Biological Uniqueness of a Ramsar Site (East Kolkata Wetlands, India); Springer: Cham, Switzerland, 2023. [Google Scholar] [CrossRef]
  2. Barua, P.; Rahman, S.H.; Eslamian, S. Coastal Zone and Wetland Ecosystem: Management Issues. In Life Below Water; Leal Filho, W., Azul, A.M., Brandli, L., Salvia, A.L., Wall, T., Eds.; Springer International Publishing: Cham, Switzerland, 2020; pp. 1–19. [Google Scholar] [CrossRef]
  3. Zhang, Z.; Huang, L.; Wang, Q.; Jiang, L.; Qi, Y.; Wang, S.; Shen, T.; Tang, B.-H.; Gu, Y. UAV Hyperspectral Remote Sensing Image Classification: A Systematic Review. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 18, 3099–3124. [Google Scholar] [CrossRef]
  4. Berhane, T.M.; Costa, H.; Lane, C.R.; Anenkhonov, O.A.; Chepinoga, V.V.; Autrey, B.C. The influence of region of interest heterogeneity on classification accuracy in wetland systems. Remote Sens. 2019, 11, 551. [Google Scholar] [CrossRef] [PubMed]
  5. Rapinel, S.; Panhelleux, L.; Gayet, G.; Vanacker, R.; Lemercier, B.; Laroche, B.; Chambaud, F.; Guelmami, A.; Hubert-Moy, L. National wetland mapping using remote-sensing-derived environmental variables, archive field data, and artificial intelligence. Heliyon 2023, 9, e13482. [Google Scholar] [CrossRef]
  6. Hejda, M.; Pyšek, P.; Jarošík, V. Impact of invasive plants on the species richness, diversity and composition of invaded communities. J. Ecol. 2009, 97, 393–403. [Google Scholar] [CrossRef]
  7. Zedler, J.B.; Kercher, S. Causes and Consequences of Invasive Plants in Wetlands: Opportunities, Opportunists, and Outcomes. Crit. Rev. Plant Sci. 2004, 23, 431–452. [Google Scholar] [CrossRef]
  8. Bradley, B.A. Remote detection of invasive plants: A review of spectral, textural and phenological approaches. Biol. Invasions 2013, 16, 1411–1425. [Google Scholar] [CrossRef]
  9. Villa, P.; Pinardi, M.; Tóth, V.; Hunter, P.; Bolpagni, R.; Bresciani, M. Remote sensing of macrophyte morphological traits: Implications for the management of shallow lakes. J. Limnol. 2016, 76, e1629. [Google Scholar] [CrossRef]
  10. Zaka, M.M.; Samat, A. Advances in Remote Sensing and Machine Learning Methods for Invasive Plants Study: A Comprehensive Review. Remote Sens. 2024, 16, 3781. [Google Scholar] [CrossRef]
  11. Joshi, C.; De Leeuw, J.; Van Duren, I.C. Remote Sensing and GIS Applications for Mapping and Spatial Modelling of Invasive Species. In Proceedings of the XXth ISPRS Congress, Istanbul, Turkey, 12–23 July 2004; ISPRS: Hanover, Germany, 2004; pp. 669–677. [Google Scholar]
  12. Yuan, S.; Liang, X.; Lin, T.; Chen, S.; Liu, R.; Wang, J.; Zhang, H.; Gong, P. A comprehensive review of remote sensing in wetland classification and mapping. arXiv 2025, arXiv:2504.10842. [Google Scholar] [CrossRef]
  13. Pang, Y.; Räsänen, A.; Wolff, F.; Tahvanainen, T.; Männikkö, M.; Aurela, M.; Korpelainen, P.; Kumpula, T.; Virtanen, T. Comparing multispectral and hyperspectral UAV data for detecting peatland vegetation patterns. Int. J. Appl. Earth Obs. Geoinf. 2024, 132, 104043. [Google Scholar] [CrossRef]
  14. Piaser, E.; Villa, P. Evaluating capabilities of machine learning algorithms for aquatic vegetation classification in temperate wetlands using multi-temporal Sentinel-2 data. Int. J. Appl. Earth Obs. Geoinf. 2023, 117, 103202. [Google Scholar] [CrossRef]
  15. Driba, D.L.; Emmanuel, E.D.; Doro, K.O. Predicting wetland soil properties using machine learning, geophysics, and soil measurement data. J. Soils Sediments 2024, 24, 2398–2415. [Google Scholar] [CrossRef]
  16. Fu, B.; Zuo, P.; Liu, M.; Lan, G.; He, H.; Lao, Z.; Zhang, Y.; Fan, D.; Gao, E. Classifying vegetation communities karst wetland synergistic use of image fusion and object-based machine learning algorithm with Jilin-1 and UAV multispectral images. Ecol. Indic. 2022, 140, 108989. [Google Scholar] [CrossRef]
  17. Scheibenreif, L.; Mommert, M.; Borth, D. Contrastive self-supervised data fusion for satellite imagery. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2022, V-3-2022, 705–711. [Google Scholar] [CrossRef]
  18. Zhang, X.; Han, L. A Generic Self-Supervised Learning (SSL) Framework for Representation Learning from Spectral–Spatial Features of Unlabeled Remote Sensing Imagery. Remote Sens. 2023, 15, 5238. [Google Scholar] [CrossRef]
  19. Grill, J.B.; Strub, F.; Altché, F.; Tallec, C.; Richemond, P.; Buchatskaya, E.; Doersch, C.; Avila Pires, B.; Guo, Z.; Gheshlaghi Azar, M.; et al. Bootstrap Your Own Latent—A New Approach to Self-Supervised Learning. In Advances in Neural Information Processing Systems; Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2020; pp. 21271–21284. Available online: https://proceedings.neurips.cc/paper_files/paper/2020/file/f3ada80d5c4ee70142b17b8192b2958e-Paper.pdf (accessed on 2 August 2025).
  20. Oquab, M.; Darcet, T.; Moutakanni, T.; Vo, H.; Szafraniec, M.; Khalidov, V.; Fernandez, P.; Haziza, D.; Massa, F.; El-Nouby, A. DINOv2: Learning Robust Visual Features without Supervision. arXiv 2023, arXiv:2304.07193. [Google Scholar] [CrossRef]
  21. Chen, X.; Xie, S.; He, K. An empirical study of training self-supervised vision transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision 2021, Montreal, QC, Canada, 10–17 October 2021; pp. 9640–9649. [Google Scholar]
  22. Chen, H.; Yokoya, N.; Chini, M. Fourier domain structural relationship analysis for unsupervised multimodal change detection. ISPRS J. Photogramm. Remote Sens. 2023, 198, 99–114. [Google Scholar] [CrossRef]
  23. Chen, W.; Zhang, Y.; Xiao, Z.; Chu, J.; Wang, X. Spectral-Spatial Self-Supervised Learning for Few-Shot Hyperspectral Image Classification. arXiv 2025, arXiv:2505.12482. [Google Scholar] [CrossRef]
  24. Liu, W.; Liu, K.; Sun, W.; Yang, G.; Ren, K.; Meng, X.; Peng, J. Self-Supervised Feature Learning Based on Spectral Masking for Hyperspectral Image Classification. Classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–15. [Google Scholar] [CrossRef]
  25. Zhu, W.; Ren, G.; Wang, J.; Wang, J.; Hu, Y.; Lin, Z.; Li, W.; Zhao, Y.; Li, S.; Wang, N. Monitoring the Invasive Plant Spartina alterniflora in Jiangsu Coastal Wetland Using MRCNN and Long-Time Series Landsat Data. Remote Sens. 2022, 14, 2630. [Google Scholar] [CrossRef]
  26. Shi, L.; Lan, X.; Duan, X.; Liu, X. FV-DLV3+: A light-weight flooded vegetation extraction method with attention-based DeepLabv3+. Int. J. Remote Sens. 2024, 46, 366–391. [Google Scholar] [CrossRef]
  27. Chen, B.; Huang, B.; Xu, B. Multi-source remotely sensed data fusion for improving land cover classification. ISPRS J. Photogramm. Remote Sens. 2017, 124, 27–39. [Google Scholar] [CrossRef]
  28. Munizaga, J.; García, M.; Ureta, F.; Novoa, V.; Rojas, O.; Rojas, C. Mapping Coastal Wetlands Using Satellite Imagery and Machine Learning in a Highly Urbanized Landscape. Sustainability 2022, 14, 5700. [Google Scholar] [CrossRef]
  29. Li, J.; Ma, J.; Ye, X. A Batch Pixel-Based Algorithm to Composite Landsat Time Series Images. Remote Sens. 2022, 14, 4252. [Google Scholar] [CrossRef]
  30. Kwan, C.; Gribben, D.; Ayhan, B.; Bernabe, S.; Plaza, A.; Selva, M. Improving land cover classification using extended multi-attribute profiles (EMAP) enhanced color, near infrared, and LiDAR data. Remote Sens. 2020, 12, 1392. [Google Scholar] [CrossRef]
  31. Reji, J.; Nidamanuri, R.R. Deep Learning-Based Multisensor Approach for Precision Agricultural Crop Classification Based on Nitrogen Levels. IEEE Geosci. Remote Sens. Lett. 2025, 22, 1–5. [Google Scholar] [CrossRef]
  32. He, C.; Wei, Y.; Guo, K.; Han, H. Removal of Mixed Noise in Hyperspectral Images Based on Subspace Representation and Nonlocal Low-Rank Tensor Decomposition. Sensors 2024, 24, 327. [Google Scholar] [CrossRef] [PubMed]
  33. Kahraman, S.; Bacher, R. A comprehensive review of hyperspectral data fusion with LiDAR and SAR data. Annu. Rev. Control. 2021, 51, 236–253. [Google Scholar] [CrossRef]
  34. Ding, G.; Liu, C.; Yin, J.; Teng, X.; Tan, Y.; He, H.; Lin, H.; Tian, L.; Cheng, J.-X. Self-supervised elimination of non-independent noise in hyperspectral imaging. Newton 2025, 1, 100195. [Google Scholar] [CrossRef] [PubMed]
  35. Chen, T.; Kornblith, S.; Norouzi, M.; Hinton, G. A Simple Framework for Contrastive Learning of Visual Representations. In Proceedings of the 37th International Conference on Machine Learning (ICML 2020), Virtual, 13–18 July 2020; PMLR: Cambridge, MA, USA, 2020; Volume 119, pp. 1597–1607. Available online: https://proceedings.mlr.press/v119/chen20j.html (accessed on 15 July 2025).
  36. Wu, B.; Hao, S.; Wang, W. Class-Aware Self-Distillation for Remote Sensing Image Scene Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 17, 2173–2188. [Google Scholar] [CrossRef]
  37. Caron, M.; Touvron, H.; Misra, I.; Jégou, H.; Mairal, J.; Bojanowski, P.; Joulin, A. Emerging Properties in Self-Supervised Vision Transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 9630–9640. [Google Scholar] [CrossRef]
  38. Ayush, K.; Uzkent, B.; Meng, C.; Tanmay, K.; Burke, M.; Lobell, D.; Ermon, S. Geography-Aware Self-Supervised Learning. Available online: https://openaccess.thecvf.com/content/ICCV2021/html/Ayush_Geography-Aware_Self-Supervised_Learning_ICCV_2021_paper.html (accessed on 27 July 2025).
  39. Mañas, O.; Lacoste, A.; Giró-I-Nieto, X.; Vazquez, D.; Rodriguez, P. Seasonal Contrast: Unsupervised Pre-Training from Uncurated Remote Sensing Data. Available online: https://openaccess.thecvf.com/content/ICCV2021/html/Manas_Seasonal_Contrast_Unsupervised_Pre-Training_From_Uncurated_Remote_Sensing_Data_ICCV_2021_paper.html (accessed on 27 July 2025).
  40. Skowronek, S.; Ewald, M.; Isermann, M.; Van De Kerchove, R.; Lenoir, J.; Aerts, R.; Warrie, J.; Hattab, T.; Honnay, O.; Schmidtlein, S.; et al. Mapping an invasive bryophyte species using hyperspectral remote sensing data. Biol. Invasions 2016, 19, 239–254. [Google Scholar] [CrossRef]
  41. Kooistra, L.; Berger, K.; Brede, B.; Graf, L.V.; Aasen, H.; Roujean, J.-L.; Machwitz, M.; Schlerf, M.; Atzberger, C.; Prikaziuk, E.; et al. Reviews and syntheses: Remotely sensed optical time series for monitoring vegetation productivity. Biogeosciences 2024, 21, 473–511. [Google Scholar] [CrossRef]
  42. Immitzer, M.; Vuolo, F.; Atzberger, C. First experience with Sentinel-2 data for crop and tree species classifications in central Europe. Remote Sens. 2016, 8, 166. [Google Scholar] [CrossRef]
  43. Tan, S.; Zhang, X.; Wang, H.; Yu, L.; Du, Y.; Yin, J.; Wu, B. A CNN-Based Self-Supervised Synthetic Aperture Radar Image Denoising Approach. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5213615. [Google Scholar] [CrossRef]
Figure 1. SSL model performance on Jilin-1, PlanetScope-3, and fused datasets using BYOL, DINO, and MoCo v3. MoCo v3 leads on single-sensor datasets, while BYOL performs best on the fused dataset.
Figure 1. SSL model performance on Jilin-1, PlanetScope-3, and fused datasets using BYOL, DINO, and MoCo v3. MoCo v3 leads on single-sensor datasets, while BYOL performs best on the fused dataset.
Plants 14 03153 g001
Figure 2. Workflow of the DeepLabV3+ fine-tuning framework. Multi-source satellite imagery is preprocessed and pretrained with BYOL, DINO, or MoCo v3, then transferred into a ResNet-50 backbone with a DeepLabV3+ decoder for segmentation map generation.
Figure 2. Workflow of the DeepLabV3+ fine-tuning framework. Multi-source satellite imagery is preprocessed and pretrained with BYOL, DINO, or MoCo v3, then transferred into a ResNet-50 backbone with a DeepLabV3+ decoder for segmentation map generation.
Plants 14 03153 g002
Figure 3. Semantic segmentation results of DeepLabV3+ fine-tuned with self-supervised pretrained backbones. (a) GT data, (b) Jilin-1 BYOL result, (c) Planet Scope 3 BYOL result, and (d) data fusion BYOL result; (e) Jilin-1 MoCo v3 result, (f) Planet Scope 3 MoCo v3 result, and (g) data fusion MoCo v3 result; (h) Jilin-1 DINO result, (i) Planet Scope 3 DINO result, and (j) Data fusion DINO result. The segmentation maps highlight diverse land cover types such as cement surfaces, buildings, pediculariids (sparse and dense), asphalt roads, open water, open land, meadows, and grassland types, with fused data often providing clearer and more accurate class boundaries than single-source imagery.
Figure 3. Semantic segmentation results of DeepLabV3+ fine-tuned with self-supervised pretrained backbones. (a) GT data, (b) Jilin-1 BYOL result, (c) Planet Scope 3 BYOL result, and (d) data fusion BYOL result; (e) Jilin-1 MoCo v3 result, (f) Planet Scope 3 MoCo v3 result, and (g) data fusion MoCo v3 result; (h) Jilin-1 DINO result, (i) Planet Scope 3 DINO result, and (j) Data fusion DINO result. The segmentation maps highlight diverse land cover types such as cement surfaces, buildings, pediculariids (sparse and dense), asphalt roads, open water, open land, meadows, and grassland types, with fused data often providing clearer and more accurate class boundaries than single-source imagery.
Plants 14 03153 g003
Figure 4. Radar chart of overall accuracy for BYOL, DINO, and MoCo v3 on Jilin-1, PlanetScope-3, and fused datasets.
Figure 4. Radar chart of overall accuracy for BYOL, DINO, and MoCo v3 on Jilin-1, PlanetScope-3, and fused datasets.
Plants 14 03153 g004
Figure 5. Location of the Bayinbuluke Alpine Wetland in Xinjiang, China, with panels showing (a) Kaidu River, Xinjiang, China, (b) Bayinbuluke Wetland, and (c) the focused study area. The legend indicates major land cover and vegetation classes used in this study.
Figure 5. Location of the Bayinbuluke Alpine Wetland in Xinjiang, China, with panels showing (a) Kaidu River, Xinjiang, China, (b) Bayinbuluke Wetland, and (c) the focused study area. The legend indicates major land cover and vegetation classes used in this study.
Plants 14 03153 g005
Figure 6. Data fusion workflow combining Jilin-1 (4-band) and PlanetScope-3 (8-band) imagery. The images are first fused using the HSI method to generate an 11-band dataset, followed by EN2N-based denoising to produce the final fused data.
Figure 6. Data fusion workflow combining Jilin-1 (4-band) and PlanetScope-3 (8-band) imagery. The images are first fused using the HSI method to generate an 11-band dataset, followed by EN2N-based denoising to produce the final fused data.
Plants 14 03153 g006
Figure 7. Workflow of data preparation, SSL feature extraction (BYOL, DINO, MoCoV3), and segmentation with DeepLabv3 for accuracy evaluation.
Figure 7. Workflow of data preparation, SSL feature extraction (BYOL, DINO, MoCoV3), and segmentation with DeepLabv3 for accuracy evaluation.
Plants 14 03153 g007
Table 1. Overall accuracy (O/A) and mean Intersection over Union (MIoU) achieved by BYOL, MoCoV3, and DINO on Jilin-1, PlanetScope-3, and fused datasets, highlighting the superior performance of MoCoV3 on fused imagery.
Table 1. Overall accuracy (O/A) and mean Intersection over Union (MIoU) achieved by BYOL, MoCoV3, and DINO on Jilin-1, PlanetScope-3, and fused datasets, highlighting the superior performance of MoCoV3 on fused imagery.
DatasetsBYOL (O/A)BYOL (mIoU)MoCoV3 (O/A)MoCoV3
(mIoU)
DINO
(O/A)
DINO (mIoU)
Jilin-193%64%72%58%76%56%
PlanetScope-392%63%86%61%84%65%
Data fusion88%64%87%63%87%57%
Table 2. Overview of datasets and supporting information used for wetland vegetation classification, including satellite imagery, topography, proximity, and reference data.
Table 2. Overview of datasets and supporting information used for wetland vegetation classification, including satellite imagery, topography, proximity, and reference data.
DataDescriptionApplication Data
Satellite ImageryHigh-resolution optical imagery acquired in June 2024 from PlanetScop-3 and Jilin-1 for wetland classification.PlanetScope-3 (9-Band Image), Jilin-1 (4-Band Image) or data-fused image (PlanetScope-3 (8-band) + Jilin-1 (4-Band) = 11-Band-Fused Image).
Topographic InformationHigh-altitude alpine basin (~2400 m elevation) with flat valley floors surrounded by the Tianshan Mountains; predominantly marshes and meadows.Alpine basin (~2400 m), whose valley floors are flat and is surrounded by Tianshan mountains; water accumulation provides the setting of marshes and alpine meadows.
Proximity InformationProximity data from field-collected GPS data and OpenStreetMap (OSM)—includes wetland boundary information, water bodies and nearby landscape features.Applied to study spatial relationships between wetlands, water bodies, and the surrounding land features; supports habitat mapping and hydrological modeling.
Reference DataReference data were generated using field-collected GPS points and OpenStreetMap (OSM) vectors. Polygons for 13 land cover and vegetation classes were manually digitized and cross-verified using high-resolution imagery. These served as ground-truth for supervised learning.Supports training and validation of deep learning models for land cover classification and segmentation; ensures label accuracy and spatial alignment across image patches.
Table 3. Distribution of ground-truth samples across 12 land cover and vegetation classes, totaling 8892 labeled instances.
Table 3. Distribution of ground-truth samples across 12 land cover and vegetation classes, totaling 8892 labeled instances.
IDClass NamesImage Samples
1Cement surface526
2Buildings668
3False out grass985
4Elymus dehurious Turoz531
5Pedicularis sparse568
6Pedicularis dense1026
7Dry meadows2522
8Leymus Scalinus200
9Asphalt road567
10Open water269
11Open land334
12Wet meadows1383
Total 8892
Table 4. Computational environment and training settings.
Table 4. Computational environment and training settings.
CategorySpecification
HardwareIntel CPU, 64 GB RAM, NVIDIA GeForce RTX 4060 (8 GB VRAM)
Operating SystemUbuntu 22.04 LTS
Python3.12.11 (Conda-Forge, MSC v.1943 64-bit)
PyTorch2.4.1 + cu124
CUDA12.4
GPU DeviceNVIDIA GeForce RTX 4060
LibrariesTorchvision 0.19.1+cu124, NumPy 1.26.4, Matplotlib 3.10.6, Lightly 1.5.20, scikit-learn 1.7.2, Rasterio 1.4.3, etc.
SoftwareArcGIS Pro
SSL Training100 epochs, batch size = 64, LR = 1 × 10−4, weight decay = 1 × 10−4
Training TimePretraining: ~1 h 45 m (±5 m), Linear evaluation: ~45 m (±5 m), Fine-tuning: ~2 h (±10 m)
Path Inference64 × 64 patches, <1 s (real time)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zaka, M.M.; Samat, A.; Abuduwaili, J.; Zhu, E.; Akhtar, A.; Li, W. Self-Supervised Learning and Multi-Sensor Fusion for Alpine Wetland Vegetation Mapping: Bayinbuluke, China. Plants 2025, 14, 3153. https://doi.org/10.3390/plants14203153

AMA Style

Zaka MM, Samat A, Abuduwaili J, Zhu E, Akhtar A, Li W. Self-Supervised Learning and Multi-Sensor Fusion for Alpine Wetland Vegetation Mapping: Bayinbuluke, China. Plants. 2025; 14(20):3153. https://doi.org/10.3390/plants14203153

Chicago/Turabian Style

Zaka, Muhammad Murtaza, Alim Samat, Jilili Abuduwaili, Enzhao Zhu, Arslan Akhtar, and Wenbo Li. 2025. "Self-Supervised Learning and Multi-Sensor Fusion for Alpine Wetland Vegetation Mapping: Bayinbuluke, China" Plants 14, no. 20: 3153. https://doi.org/10.3390/plants14203153

APA Style

Zaka, M. M., Samat, A., Abuduwaili, J., Zhu, E., Akhtar, A., & Li, W. (2025). Self-Supervised Learning and Multi-Sensor Fusion for Alpine Wetland Vegetation Mapping: Bayinbuluke, China. Plants, 14(20), 3153. https://doi.org/10.3390/plants14203153

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop