Next Article in Journal
Efficient On-Board Compression for Arbitrary-Shaped Cloud-Covered Remote Sensing Images via Adaptive Filling and Controllable Quantization
Previous Article in Journal
Mapping Natural Populus euphratica Forests in the Mainstream of the Tarim River Using Spaceborne Imagery and Google Earth Engine
Previous Article in Special Issue
Finding Oasis Cold Island Footprints Based on a Logistic Model—A Case Study in the Ejina Oasis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Snow Cover Extraction from Landsat 8 OLI Based on Deep Learning with Cross-Scale Edge-Aware and Attention Mechanism

1
College of Urban and Environmental Science, Northwest University, Xi’an 710127, China
2
Shaanxi Key Laboratory of Earth Surface System and Environmental Carrying Capacity, Northwest University, Xi’an 710127, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2024, 16(18), 3430; https://doi.org/10.3390/rs16183430
Submission received: 11 July 2024 / Revised: 4 September 2024 / Accepted: 14 September 2024 / Published: 15 September 2024

Abstract

:
Snow cover distribution is of great significance for climate change and water resource management. Current deep learning-based methods for extracting snow cover from remote sensing images face challenges such as insufficient local detail awareness and inadequate utilization of global semantic information. In this study, a snow cover extraction algorithm integrating cross-scale edge perception and an attention mechanism on the U-net model architecture is proposed. The cross-scale edge perception module replaces the original jump connection of U-net, enhances the low-level image features by introducing edge detection on the shallow feature scale, and enhances the detail perception via branch separation and fusion features on the deep feature scale. Meanwhile, parallel channel and spatial attention mechanisms are introduced in the model encoding stage to adaptively enhance the model’s attention to key features and improve the efficiency of utilizing global semantic information. The method was evaluated on the publicly available CSWV_S6 optical remote sensing dataset, and the accuracy of 98.14% indicates that the method has significant advantages over existing methods. Snow extraction from Landsat 8 OLI images of the upper reaches of the Irtysh River was achieved with satisfactory accuracy rates of 95.57% (using two, three, and four bands) and 96.65% (using two, three, four, and six bands), indicating its strong potential for automated snow cover extraction over larger areas.

1. Introduction

Snow cover, as a crucial component of the cryosphere, is highly sensitive to climatic conditions and serves as a significant indicator of climate change [1,2]. It regulates the surface energy balance through the albedo effect. As snow cover diminishes, increased solar radiation absorption leads to localized warming, thereby accelerating the process of climate warming. Additionally, the retreat of snow cover may trigger the melting of permafrost, releasing substantial amounts of greenhouse gases and further advancing global climate change. In mountainous basins, snow cover is a vital water source, significantly influencing river runoff and freshwater supply [3]. Snowmelt serves as a crucial replenishment for many rivers, especially in arid or semi-arid regions, where the melting of snow delivers meltwater downstream, extending the water supply throughout spring and summer, which is essential for agricultural irrigation, urban water supply, and hydropower generation. However, as the climate warms and transitions from snow to rain in colder regions, the timing of snowmelt shifts earlier, reducing solid water storage in winter, and snowmelt and water flows move earlier in the season. This change can lead to increased risk of flooding in spring and exacerbate water scarcity in summer [4]. Therefore, understanding the distribution, storage, and melting of snow in mountainous regions with complex topography is crucial for effective water resource management, maintaining surface energy balance, and addressing regional climate change impacts [5,6,7].
Currently, snow cover analysis primarily depends on remote sensing data [8]. The evolution of snow cover monitoring algorithms has progressed from traditional visual interpretation to machine learning and deep learning [6,9,10]. Meanwhile, combining multiple data sources has led to significant progress in remote sensing data, particularly in temporal and spatial resolution [11]. However, when using optical remote sensing images for snow cover monitoring, the accuracy of snow cover identification tends to decrease as cloud coverage increases in some areas [12]. In addition, terrain has a significant impact on the distribution and accumulation of snow cover [13]. Mountains, canyons, and valleys can lead to shadow occlusion in remote sensing images [14]. The intermingling of different land cover types (e.g., snow cover, vegetation, rocks, forests) can lead to a complex blend of reflected signals [15]. Therefore, traditional thresholding methods and machine learning methods cannot fully extract useful information, making snow extraction still challenging [16], with various deep learning algorithms being increasingly applied.
In the realm of remote sensing for snow cover, mainstream deep learning algorithms primarily deploy architectures such as U-net [17] and DeepLabV3+ [18], often in conjunction with Transformer [19] architectures to boost performance. For instance, Wang et al. [20] assessed the U-net model’s performance in mapping snow cover using various Sentinel-2 image bands and discovered that bands B2, B4, B11, and B9 preserved almost all pertinent information for the clouds, snow, and background. They highlighted the challenges in differentiating snow from clouds, noting that resolving these complexities necessitates sophisticated models and superior data quality. Xing et al. [21] introduced a Partial Convolution U-Net (PU-Net) that leverages spatial and temporal data to reconstruct obscured NDSI values in MODIS snow products. The reconstructed pixels displayed Mean Absolute Errors ranging from 4.22% to 18.81% and Coefficients of Determination between 0.76 and 0.94 across various NDSI regions and coverage scenarios, underscoring the U-net model’s capability in snow detection via remote sensing. Nevertheless, Yin et al. [22], through comparative analyses involving threshold segmentation, U-net, Deeplabv3+, CDUnet, and their novel Unet3+ methods in cloud–snow discrimination trials, demonstrated that while U-Net and its enhancements surpass traditional approaches in accuracy, they still suffer from notable deficiencies such as inadequate long-range dependency and restricted receptive fields.
Simultaneously, facing challenges such as the multiscale features of snow in high-resolution remote sensing images, the similarity between clouds and snow, and occlusions caused by mountains and cloud shadows, Guo et al. [23] noted the scarcity and labor-intensive nature of obtaining pixel-level annotations for snow mapping in HSRRS images. They pre-emptively applied transfer learning techniques with the DeepLabv3+ network, achieving a snow extraction accuracy of 91.5% MIoU on a limited dataset, surpassing the 81.0% accuracy of a comparable FCN model. Kan et al. [24] introduced an edge enhancement module in the DSRSS-Net model, optimizing feature affinity loss and edge loss, which resulted in enhanced detailed structural information and improved edge segmentation for cloud–snow segmentation tasks. Relative to MOD10A1, the DSRSS-Net model’s snow classification accuracy and overall precision increased by 4.45% and 5.1%, respectively. Wang et al. [25] incorporated a Conditional Random Field (CRF) model into the DeepLabV3+ framework to overcome its limitations in distinguishing clouds and snow in GF-1 WFV imagery due to the absence of a short-wave infrared band sensor, noting a significant reduction in boundary blurring, slice traces, and isolated misclassifications. However, the DeepLabV3+ network’s complex structure contributes to its high computational demand. Additionally, Ding et al. [26] proposed the MAINet method, combining CNN and Transformer architectures in the final downsampling stage to minimize detail loss and clarify deep semantic features by reducing redundancy. Similarly, Ma et al. [27] used a dual-stream structure to merge information from Transformer and CNN branches, enhancing the model’s ability to differentiate between snow and clouds with similar spectral features during snow cover mapping using Sentinel-2 imagery, though further exploration of the model’s performance in practical applications is needed.
In summary, although different deep learning algorithms have been used for snow cover extraction, there is still great potential for improvement in terms of encoder–decoder structure and image structure information utilization, etc. Therefore, this study proposes a snow cover extraction method based on cross-scale edge-aware fusion and a channel-space attention mechanism in parallel with U-net. Section 2 offers a detailed overview of the two types of datasets used for experimental evaluation, as well as the specific methodologies employed in the models’ construction and the experimental environments and parameter configurations for each model. Section 3 displays the experimental results of the models across various datasets from multiple angles, as well as the ablation study outcomes of the CEFCSAU-net model. Section 4 visualizes and deeply analyzes the features extracted from the CSA and CEF modules, examines the limitations of the models discussed in this paper in applying multi-source remote sensing data, and outlines future research strategies and directions. Finally, Section 5 summarizes the application prospects and principal findings of the experimental results of the models discussed in this paper.

2. Materials and Methods

2.1. Experimental Dataset

The datasets evaluated in this study include the publicly available CSWV_S6 dataset and Landsat8 OLI dataset, respectively.

2.1.1. CSWV_S6 Dataset

The CSWV dataset [28] is a freely and publicly available dataset constructed by Zhang et al., based on synthesizing WorldView2’s red, green, and blue bands. The dataset was taken in the Cordillera Mountains of North America, and the distribution of clouds and snow was obtained based on the interpretation of WorldView2 images from June 2014 to July 2016. This study utilizes the CSWV_S6 data with a spatial resolution of 0.5 m (Figure 1), available at https://github.com/zhanggb1997/CSDNet-CSWV (accessed on 14 September 2024). The area depicted in the image comprises forests, grasslands, lakes, and bare ground, among others. In the dataset, clouds and snow were labeled, but only the snow labels were used in this study to classify features into snow and non-snow categories.

2.1.2. Landsat 8 OLI Dataset

Landsat 8 OLI imagery offers a spatial resolution of 30 m, with the OLI sensor comprising nine bands, including three visible, infrared, and short-wave infrared bands, among others. To minimize cloud interference during snow extraction, we selected six Landsat images with cloud coverage of less than 30% from dates “20180413”, “20180515”, “20190315”, “20190603”, “20200317”, and “20200402”. Figure 2 presents an overview of the study area, with complete data available at https://zenodo.org/records/12635483 (accessed on 14 September 2024). In the methodological experimentation phase, bands 2 (blue), 3 (green), and 4 (red) were selected to construct the dataset for comparative analysis against the public dataset (CSWV_S6). For labeling, bands 3 (green), 5 (NIR), and 6 (SWIR 1) were utilized. Additionally, recognizing the significance of the short-wave infrared band for snow detection, the study incorporated SWIR 1 (band 6) as an input feature in the discussion phase experiments. This was done to compare the efficacy of solely using bands 2, 3, and 5 and to delve deeper into the benefits of multispectral data fusion for improving the performance of remote sensing image processing algorithms.

2.2. Data Processing

First, radiometric calibration and atmospheric correction were performed on Landsat 8 OLI image data. Snow cover was then extracted using the SNOMAP algorithm by calculating the Normalized Difference Snow Index (NDSI) with the short-wave infrared and green bands of the Landsat images [29]. In experiments, pixels were extracted as snow cover if they had an NDSI of 0.4 or higher, and the near-infrared reflectance threshold was increased to 0.11 or higher to reduce the misclassification of water bodies. Additionally, pixels that were either not identified or misidentified by the SNOMAP algorithm were corrected through visual interpretation to accurately determine the snow cover extent within the study area as label data.
Then, due to the limitations in computational resources and model architecture when processing large-scale data, large-size remote sensing images were divided into multiple fixed-size chunks [30]. Therefore, the study utilized the sliding window method to segment the six-image views of CSWV_S6 data, Landsat 8 OLI data, and the corresponding labeled data to create an input dataset that meets the model’s required size. Here, a fixed-size sliding window of 512 × 512 pixels was established, and the sliding window moved along the horizontal and vertical directions on each remote sensing image, advancing a fixed step each time. In this manner, the entire remote sensing image was gradually covered. The CSWV_S6 image data comprised 1114 windows of 512 × 512 size, while the Landsat 8 OLI image data included 1260 windows of the same size, with corresponding labeling data for each.
The homemade experimental data underwent a secondary screening, with datasets containing 10% to 90% snow pixels selected to focus the study on snow-covered scenes. Ultimately, from the 1260 original window datasets of Landsat data, 413 that met the criteria were selected. For the 1114 raw window datasets of CSWV_S6 data, to minimize screening interference and emphasize generalization capabilities, snow-free datasets were excluded, resulting in 688 valid datasets being selected. This approach not only ensures the relevance of the data to the study’s objectives but also enhances the robustness of the findings by covering a wide range of snow conditions.
Following the data preprocessing outlined above, the labels and individual band data were stored in single-channel .tif format, with dimensions of (512, 512, 1). To boost data processing efficiency, streamline the preprocessing workflow, and enhance compatibility with deep learning frameworks, the data in .tif format were converted to .npy format. Additionally, all image data were normalized, scaling pixel values to the range [0,1]. The converted label data retained its shape (512, 512, 1), while the Landsat 8 OLI data were stored in the following multi-channel formats: (512, 512, 3) and (512, 512, 4). Similarly, the CSWV_S6 data were converted into a (512, 512, 3) .npy format. This conversion step ensured the data were efficiently processed and readily usable in subsequent deep learning models.
Finally, the valid data were divided into training, validation, and test sets in the ratio of 7:2:1. Additionally, for deep learning model training, sample size is typically a critical consideration. To ensure that the deep learning model could learn a sufficient number of features and patterns, this experiment applied horizontal and vertical flips to the data inputs with a 50% probability during each cycle of the training and validation phases. This technique enhanced the robustness of the model by augmenting the variability of training examples without requiring additional raw data.

2.3. Deep Learning Models

2.3.1. Model Architecture

We propose a cross-scale edge-aware fusion and channel spatial attention mechanism parallelization-based snow cover extraction method on the U-net model architecture (Cross-Scale Edge-Aware Fusion and Channel Spatial Attention Mechanism Parallelization, CEFCSAU-net) (Figure 3). First, an enhanced parallel Channel Spatial Attention Mechanism (CSA) is employed during the encoding stage of model feature extraction to adaptively enhance the model’s attention to key features and improve the efficiency of global semantic information utilization [31]. Second, the proposed Cross-Scale Edge-Aware Fusion (CEF) module replaces the original skip connections in U-net, enhancing low-level image features through multiple edge detection at shallow scales and improving detail perception through the separation and fusion of features at deeper scales [32]. Additionally, Batch Normalization (BN) is applied after each convolutional stage to accelerate training and stabilize the gradients [33].
Additionally, the model’s output is provided as “logits”, which facilitate the calculation of the loss function. These logits represent raw, unbounded predictive values, produced directly by the final layer of the model without applying an activation function. They reflect the model’s confidence in each pixel belonging to a specific category. In binary classification tasks, each pixel’s final classification is determined by applying a zero threshold to these logits. Specifically, if the value of logits exceeds zero, the pixel is classified as the target class (snow); otherwise, it is classified as the background or other categories (non-snow). This approach enables the model to classify effectively based on the sign and magnitude of the logits, eliminating the need for additional nonlinear transformations. Ultimately, following this post-processing, the output data from the test set are stored in PNG format with dimensions of (512, 512, 1).

2.3.2. Channel Spatial Attention Mechanism Parallel Module (CSA)

The parallel channel space attention mechanism modules (Figure 4) are introduced during the encoding phase of model feature extraction, which adaptively enhance the model’s attention to key features and improve the efficiency of utilizing global semantic information.
The channel attention mechanism [34] in the upper branch of the CAS module primarily concentrates on the weights assigned to different channels in the input feature map, aiding the model in better learning the correlations among features and enhancing the model’s representational capacity.
The spatial attention mechanism in the lower branch primarily concentrates on the weights assigned to different locations in the input feature map, thereby honing the model’s focus on crucial spatial locations and enhancing its capacity to perceive spatial structures [35]. This combined approach enhances the model’s capacity to process input features, thereby boosting its performance and generalization capabilities. The calculation formula is as follows:
C A f = X Softmax MLP AvgPool   X
Sf = Con v 1 × 1 Concat MaxPool X   ,   AvgPool X
S A f = X Softmax Con v 1 × 1 Sf
CS A f = Con v 1 × 1 Concat C A f , S A f

2.3.3. Cross-Scale Edge-Aware Fusion Module (CEF)

The cross-scale edge-aware fusion module (Figure 5) replaces the traditional skip connections in the U-net model. The aim is to enhance low-level image features by introducing multiple edge detection on shallow feature scales and to improve detail perception through branch separation and fusion of features on deep feature scales.
In the CEF module, image feature information is initially captured on shallow feature scales using Sobelx, Sobely, and Laplacian operations [36]. Among them, Sobelx and Sobely operations detect horizontal and vertical edges, while the Laplacian operation enhances high-frequency details [24]. These operations are processed in parallel and combined to enhance the model’s sensitivity to image edges and textures, as well as to improve feature differentiation. The computation and output equations are as follows:
X = Con v 1 × 1 X
SS L f = Conv X , kerne l x + Conv X , kerne l y + Conv X , kerne l L
kerne l x = 1 0 1 2 0 2 1 0 1 ,   kerne l y = 1 2 1 0 0 0 1 2 1 ,   kerne l L = 0 1 0 1 4 1 0 1 0
where Conv denotes the convolution operation, kernelx, kernely, and kernelL denote the convolution kernels for Sobelx, Sobely, and Laplacian operations, respectively, and SSLf is the new feature map after their summation. The new feature map is then used to generate the edge-sensing features through a series of operations: the edge-sensing feature (Esf).
DF = Upsample RELU BN Con v 1 × 1 Y f
C s f = Con v 3 × 3 X + RELU BN Con v 1 × 1 X + DF
D f = X DF
Ou t f = E s f + C s f + D f

2.4. Snow Extraction

2.4.1. Experimental Configuration

Before conducting experiments with the research model described in this paper, establishing the appropriate experimental environment is essential. Currently, the software environment for the model in this paper is configured with the “Windows 11, version 22H2” operating system, “Python 3.8” version, and the “PyTorch 1.7.1+cu110” deep learning framework. The hardware environment features a “CPU: 12th Gen Intel Core i7-12700K” and a “GPU: NVIDIA GeForce RTX 3090”.
During the model’s construction and training phases, it was trained on a GPU device, with the number of training cycles (epochs) uniformly set at 50, batch size at 8, and worker threads in the data loader also at 8. In each training cycle, BCEWithLogitsLoss [37] was used as the loss function, with its formula provided in Equations (12) and (13).
σ x i = Sigmoid x = 1 1 + e x i
BCE W Loos = 1 N i = 1 N y i log σ x i + 1 y i log 1 σ x i
where N is the number of samples; yi is the true label of sample i, which takes the value of 0 or 1; xi is the prediction result of sample i; and the log function is the natural logarithm.
In addition, the experiment employs the Adam optimizer for parameter updates, with an initial learning rate (α) of 0.0001, a first-order momentum decay rate (β1) of 0.5, and a second-order momentum decay rate (β2) of 0.99, facilitating faster convergence and enhanced performance of the model.
To compare the performance of different models under uniform experimental conditions, enhancing comparability and consistency, this paper standardized the hyperparameter settings of the FCN8s [38], SegNet [39], U-net [17], and DeepLabV3+ [18] models. Specifically, FCN8s based on the VGG16 architecture employ a fully convolutional network (FCN) that integrates features from various levels via skip connections. SegNet features a classic encoder–decoder architecture, with the encoder identical to VGG16 and the decoder progressively upsampling to restore spatial resolution. U-net is known for its symmetrical U-shaped structure, which utilizes skip connections to directly transfer features from the encoder to the decoder. DeepLabV3+ employs ResNet50 as its backbone network, expanding the receptive field by incorporating atrous convolution and an ASPP module, capturing multi-scale features.

2.4.2. Assessment of Indicators

To objectively evaluate the prediction accuracy of the model at the image pixel level and its ability to recognize snow categories, four metrics were selected to evaluate the model’s performance in this study: Pixel Accuracy (PA), Recall (R), F1 Score (F1), and Mean Intersection over Union (MIoU) [28]. The formulas for these metrics are as follows:
P A = T P + T N T P + T N + F P + F N
P = TP TP + FP
R = T P T P + F N
F 1 = 2 P * R P + R
I o U i = T P i T P i + F P i + F N i
M I o U = 1 N i = 1 N Io U i
where PA is the ratio of the number of pixels correctly categorized at all pixel levels to the total number of pixels, TP denotes true cases, TN denotes true-negative cases, FP denotes false-positive cases, and FN denotes false-negative cases. IoUi denotes the intersection and concatenation ratio of the ith category, and N denotes the total number of categories.

3. Experimental Results

3.1. Experimental Results for CSWV_S6 Data

Table 1 presents the average results for all evaluation metrics in the test set. It is evident that the models differ in terms of pixel accuracy (PA), recall (R), F1 score, and mean intersection over union (mIoU). The CEFCSAU-net model outperforms other models across all indices, particularly in pixel accuracy, achieving 98.14% and demonstrating high segmentation accuracy and consistency on CSWV_S6 data.
Simultaneously, three images from the CSWV_S6 test set, featuring snow cover ranging from 10% to 90%, were selected (Figure 6). The results indicate that among the FCN8s, SegNet, U-net, DeepLabV3+, and CEFCSAU-net models, the CEFCSAU-net exhibits the fewest segmentation errors.

3.2. Experimental Results on Landsat8 OLI Data

Table 2 displays the performance of each model on the snow cover extraction task, based on experiments with various segmentation models on the self-created Landsat 8 OLI dataset. The experimental results reveal that the CEFCSAU-net model achieves the highest scores across all metrics, underscoring its superior performance in snow extraction. Combining some of the test samples (Figure 7) shows that the overall performance results of each model align with those of the CSWV_S6 data.
Furthermore, in scenes with high cloud content, like those in the fifth row, the FCN8s, SegNet, and U-net models frequently misclassify significant amounts of clouds as snow. In contrast, the DeepLabV3+ and CEFCSAU-net models effectively identify clouds as non-snow features. The CEFCSAU-net model, however, retains a keen perception of snow edges, reflecting its high accuracy and robustness in snow extraction. This analysis illustrates the relative strengths and weaknesses of the different models under varied environmental conditions.

3.3. Model Comparison and Analysis

Several studies have carried out snow cover monitoring with pixel-level convolutional neural network algorithms [23,40,41], finding that their powerful feature learning capabilities, adaptability to complex textures, and processing of temporal information lead to more accurate, automated, and efficient monitoring and analysis of snow cover [42]. In conjunction with Section 3.1 and Section 3.2, we compare and analyze several models selected in the experiment.
The experimental findings are as follows: the FCN8s model, featuring a full convolutional network structure, fails to fully utilize edge detail information, resulting in poor edge effects and a high number of erroneous detections. It also performs poorly across various evaluation metrics. The SegNet and U-net models, which utilize an encoder–decoder structure, improve the retention of edge detail information. However, they still lack spatial information, resulting in fewer edge detail misdetections compared to the FCN8s model, but with more widespread misdetection areas.
Meanwhile, the DeepLabV3+ model, which contains pyramid pooling and atrous convolution structures, achieves multi-scale feature extraction and integration. This, combined with fusion with low-level feature maps, restores the spatial details of the segmentation results and performs more stably across all evaluation metrics.
Finally, compared to the other four models, the CEFCSAU-net model introduces the CSA module, which adaptively enhances the model’s focus on key features and improves the utilization of global semantic information. Additionally, the original skip-connected CFE module of U-net is replaced to enhance cloud–snow differentiation and edge perception through multiple edge detection and branch separation and fusion features, ultimately achieving the best performance results.

3.4. Indicator Fluctuations Due to Data Discrepancies

From the experimental results of the previous CEFCSAU-net model on CSWV_S6 data and Landsat8 OLI data, the indicators of PA, R, F1, and mIoU are higher than other models. Furthermore, the results of the ablation experiments confirm that the incremental integration of the CSA and CEF modules into the U-net model effectively enhances its performance. Meanwhile, by comparing the boxplots of various indicators of different models on the two sets of data in Figure 8, it can be found that the indicator scores of the CEFCSAU-net model for the whole test set of data are stable, compact, and uniformly distributed compared with the other four models. This further confirms the significant role of the improvement measures discussed in this paper in enhancing model performance.
However, as depicted in Figure 8, the fluctuation range for the R, F1, and mIoU metrics in Figure 8a is substantially greater than for the PA metrics, with more outliers observed for each model than in Figure 8b. Analysis of each model’s results in the test set revealed that the metric fluctuations stemmed from data variations, notably because the CSWV_S6 dataset excluded no-snow data during processing, leading to a higher presence of data with a lower percentage of snow image elements. The scores of the evaluation metrics for each model on three CSWV_S6 test set examples are displayed in Figure 9, illustrating how this type of data induces minor errors in recognizing snow image elements, thereby causing significant fluctuations in the scores of the R, F1, and mIoU metrics.
Nevertheless, the performance of CEFCSAU-net in Figure 9 is still satisfactory. It shows that the study’s data processing of CSWV_S6 did not lead to the emergence of the category imbalance problem, but rather verified the reliability of the CEFCSAU-net model.

3.5. Differences in the Extraction of Cloud–Snow Confusion Scenarios

As is well known, distinguishing between snow, clouds, and cloud shadows is a crucial step in remote sensing processing [23]. Yao et al. [43] introduced the CD-AttDLV3+ network, which employs channel attention and demonstrates outstanding performance in distinguishing clouds from bright surfaces and detecting translucent thin clouds, providing valuable insights. In Section 3.1 and Section 3.2, Figure 6 and Figure 7 illustrate that the CEFCSAU-net model has fewer error detection regions than the other models, both on the edges and in the details. However, the CEFCSAU-net model, while generally stable, shows an increased red error detection region in scenes with high cloud coverage, as seen in the fifth row of Figure 7. Analysis of the input image reveals that this error region primarily originates from the overlap of clouds and snow.
In addition, while the CDnet proposed by Yang et al. introduces edge refinement operations, it encounters pixel omission issues in thin cloud extraction [44]. Figure 10 shows that the CEFCSAU-net model demonstrates capabilities in detecting snow under thin clouds and effectively distinguishing large areas of thick clouds, though it sometimes struggles with the boundaries of thick clouds. This indicates that even the high-performing CEFCSAU-net model experiences increased false detections in areas with cloud–snow confusion.

3.6. Ablation Experiments

In deep learning modeling research, ablation experiments quantify the importance of specific components, hierarchies, parameters, or input features of a model by gradually modifying or removing them on model functionality, performance, or robustness [45]. In this experiment, ablation experiments were conducted for both the CSA module and the CEF module on both the CSWV_S6 dataset and the Landsat 8 OLI dataset in order to observe and compare the performance under different model configurations, and the experimental correlation results are shown in Figure 11 below.
From the perspective of model design principles and architecture, the parallel channel and spatial attention mechanisms focus on the correlations among different channels and the spatial connections among various locations in the input feature map, respectively, in order to improve the model’s representation ability and perception of spatial structure, as well as to improve the model’s processing of the input features and thus the accuracy of classification. The experimental results show that the introduction of the CSA module on the CSWV_S6 dataset improves the PA, R, F1, and mIoU by 1.12%, 9.49%, 9.88%, and 5.8%, respectively. On the Landsat 8 OLI dataset, these metrics improved by 2.74%, 1.12%, 2.66%, and 4.87%, respectively, demonstrating the CSA module’s effectiveness.
Additionally, low-level image features are enhanced by introducing multiple edge detections on the shallow feature scale, as well as the cross-scale CEF module, which obtains detailed features through feature fusion and separation after upsampling on the deep feature scale. Following its introduction, the CEF module also improved various evaluation metrics on the CSWV dataset: PA by 1.16%, R by 15.21%, F1 by 8.78%, and mIoU by 5.29% compared to the U-net model. On the Landsat 8 OLI dataset, improvements were PA by 1.41%, R by 1.7%, F1 by 1.63%, and mIoU by 1.64%, indicating the CEF module’s effectiveness.
Finally, the CEFCSAU-net model, which achieved optimal results and significant enhancements across all metrics, demonstrates that when both the CSA and CEF modules are introduced, the features enhanced by the CSA module are further enhanced by the cross-scale CEF module. This synergy results in improved outcomes with reduced information loss in differential features.

4. Discussion

4.1. Contribution Analysis of CSA and CEF Modules

To gain a deep understanding of the roles and contributions of the CSA and CEF modules in the feature extraction process, this paper visualized the intermediate features extracted by these modules as heatmaps to facilitate an intuitive observation of the important information captured by these modules during data processing [46].
CSA Module Contribution Analysis: The CSA module enhances key feature expressions by focusing on both the channel and spatial dimensions of the feature maps. Considering the number of channels in the intermediate feature maps, this study visualized only the first eight channels of the feature maps before and after the first application of the CSA module (total number of channels: 64). In Figure 12b, it is evident that half of the channel feature maps were not effectively distinguished, while the other half showed only subtle differences. However, in Figure 12c, each channel’s feature map is effectively distinguished, and significant differences are observed between the channels. These results demonstrate how the CSA module’s channel attention mechanism enhances and suppresses various channels and how its spatial attention mechanism discriminates the information across feature maps. This selective attention allows the model to utilize key features more accurately in subsequent layers, thereby enhancing overall performance.
CEF Module Contribution Analysis: The CEF module captures edge features and enhances detail perception by extracting cross-scale features. In experiments, with Figure 12a as input and by manipulating the presence of the CSA and CEF modules, the value of the CEF module is assessed. In Figure 13, it is evident that the feature maps b1–b4 for snow and non-snow edges are significantly enhanced compared to a1–a4, demonstrating the model’s sustained sensitivity to spatial features with scale changes. This illustrates the CEF module’s strong structural capability in spatially enabling the model to more effectively capture the geometric shapes and contours of key features in the input data.
Combined Contribution Analysis and Limitations of CSA and CEF Modules: As observed, feature maps c1 and c2 in Figure 13 offer superior performance in detailing edges and spatial attributes compared to a1, a2, b1, and b2, providing the model with a more comprehensive capability for feature expression. However, feature maps at stages c3 and c4 show less spatial detail in the upper right areas than a3, a4, b3, and b4. This may be attributed to the feature maps being represented as channel averages, which do not effectively showcase the distinct contributions of each channel in a collaborative state. Additionally, the abundance of channels in the network’s deeper stages and the successive normalizations may reduce feature value distinctions, diminishing module performance.
In conclusion, the CSA and CEF modules play an indispensable role in the feature extraction process. For a more detailed analysis of the value and role of different modules within the model, considering the mechanics of remote sensing imagery and integrating multispectral data for further targeted, quantitative analysis and exploration are necessary.

4.2. Limitation and Prospects

The snow cover extraction method described in this paper, based on cross-scale edge perception and attention mechanisms, is implemented on the U-net model architecture. This approach addresses the issues of inadequate local detail perception and insufficient global semantic information utilization in existing methods by incorporating the CSA and CEF modules. Furthermore, it demonstrates potential in detecting snow under thin clouds. However, due to the limitations inherent in optical remote sensing images, the current CEFCSAU-net model still faces challenges with cloud–snow confusion scenarios, revealing some limitations and shortcomings.
Additionally, in traditional snow detection methods, the SWIR band of the NDSI plays a crucial role [47]. Consequently, this study specifically incorporated the SWIR band in analyzing Landsat data to assess its impact on the performance of deep learning models. The experimental results are shown in Table 3. Comparison with Table 2 reveals that the overall performance of the SegNet and CEFCSAU-net models has improved, particularly the CEFCSAU-net model, which exhibited exceptional performance. However, some metrics of other models demonstrated a decline in performance, possibly due to the minimal differences in snow across visible light bands (such as blue, green, red), which led to certain limitations in model performance due to information redundancy. These results indicate that the CEFCSAU-net model exhibits consistent performance; however, further research is needed to explore how multispectral data fusion could potentially enhance remote sensing image processing algorithms.
In the future, experimental validation using larger and richer and more diverse datasets will be considered to improve the generalization ability and robustness of the model, as well as to explore the generalization ability of the model across datasets, verify the performance of the model on different datasets, etc. Additionally, there is interest in developing interpolated cloud-increasing, wetting, and drying strategies to derive robust SAR training data, leveraging the cloud-penetrating capabilities of synthetic aperture radar (SAR) and the concept of interpolating de-clouded data using multi-temporal phase data from optical sensors [48]. Among them, the cloud-increasing strategy is based on the consideration that although SAR has the ability to penetrate the cloud layer, the cloud layer may introduce some errors or interference in the radiation process. When SAR penetrates clouds, it may capture cloud reflection or scattering signals, potentially interfering with the extraction and decoding of the surface signals. The humidification and drying strategies then serve as replacement strategies for image elements with ambiguous snow states. The labeling of SAR data also involves using contemporaneous optical images.
The envisioned approach aims to construct a deep learning dataset based on SAR data; however, it must be acknowledged that dataset construction will involve data from various satellites on different scales, given current technological limitations. Therefore, further model design and construction should consider cross-scale considerations and potentially involve innovations in multi-task loss function design [24].

5. Conclusions

This paper tackles the challenges of insufficient local detail perception and inadequate global semantic information utilization in existing deep learning methods for remote sensing image snow cover extraction by proposing a pixel-level semantic segmentation network based on the U-net architecture CEFCSAU-net. By integrating parallel channel and spatial attention (CSA) mechanisms and a cross-scale edge-aware feature fusion (CEF) module, this model significantly enhances its ability to capture edges, spatial details, and contextual information. This provides a more accurate and robust tool for snow cover detection, which is crucial in studies of climate change and hydrological processes. Additionally, the robust adaptability exhibited by the CEFCSAU-net model indicates its broad applicability not only in snow detection but also in remote sensing tasks like land cover classification, urban mapping, and disaster response monitoring.
Experimental results demonstrate that the CSA module of CEFCSAU-net can effectively focus on the correlations among different channels and the spatial connections within the input feature maps, while the CEF module offers superior capabilities for edge detail extraction. Across multiple evaluation metrics on the Landsat 8 OLI and the CSWV_S6 public optical remote sensing datasets, CEFCSAU-net outperforms existing methods. Specifically, the accuracy on the CSWV_S6 dataset is 98.14%, and on the Landsat 8 OLI images, it achieves 95.57% (using bands 2, 3, and 4) and 96.65% (using bands 2, 3, 4, and 6). These results underline that the introduction of the SWIR band not only enhances model performance but also offers a promising avenue for further exploration into multispectral data fusion.

Author Contributions

Conceptualization, Z.Y. and S.Z.; methodology, Z.Y., H.G. and S.Z.; software, Z.Y. and H.G.; validation, Z.Y., H.G., S.Z. and W.W.; formal analysis, Z.Y. and S.Z.; investigation, Z.Y. and W.W.; resources, S.Z.; data curation, H.G.; writing—original draft preparation, Z.Y., H.G., S.Z. and W.W.; writing—review and editing, Z.Y., H.G., S.Z. and W.W.; visualization, Z.Y.; supervision, Z.Y. and S.Z.; project administration, S.Z.; funding acquisition, S.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received support from the General Program of the China National Natural Science Foundation (Grants No42171124) and the Key Program of China National Natural Science Foundation (Grants No42330512).

Data Availability Statement

The data of CSWV_S6 and Landsat 8 OLI used to support this study are publicly available. The CSWV_S6 data can be downloaded from the website https://github.com/zhanggb1997/CSDNet-CSWV (accessed on 14 September 2024). The Landsat 8 OLI data can be downloaded from the website https://zenodo.org/records/12635483 (accessed on 14 September 2024). The code is available upon request.

Acknowledgments

Thanks to all the anonymous reviewers for their constructive and valuable suggestions on the earlier drafts of this manuscript.

Conflicts of Interest

The authors declare that there is no conflict of interest regarding the publication of this article.

References

  1. Romanov, P. Global multisensor automated satellite-based snow and ice mapping system (GMASI) for cryosphere monitoring. Remote Sens. Environ. 2017, 196, 42–55. [Google Scholar] [CrossRef]
  2. Ding, A.; Jiao, Z.; Zhang, X.; Dong, Y.; Kokhanovsky, A.A.; Guo, J.; Jiang, H. A practical approach to improve the MODIS MCD43A products in snow-covered areas. J. Remote Sens. 2023, 3, 0057. [Google Scholar] [CrossRef]
  3. Musselman, K.N.; Addor, N.; Vano, J.A.; Molotch, N.P. Winter melt trends portend widespread declines in snow water resources. Nat. Clim. Change 2021, 11, 418–424. [Google Scholar] [CrossRef] [PubMed]
  4. Han, J.; Liu, Z.; Woods, R.; McVicar, T.R.; Yang, D.; Wang, T.; Yang, Y. Streamflow seasonality in a snow-dwindling world. Nature 2024, 629, 1075–1081. [Google Scholar] [CrossRef]
  5. Banerjee, A.; Chen, R.; Meadows, M.E.; Sengupta, D.; Pathak, S.; Xia, Z.; Mal, S. Tracking 21st century climate dynamics of the Third Pole: An analysis of topo-climate impacts on snow cover in the central Himalaya using Google Earth Engine. Int. J. Appl. Earth Obs. Geoinf. 2021, 103, 102490. [Google Scholar] [CrossRef]
  6. Ma, Y.; Huang, X.D.; Yang, X.L.; Li, Y.X.; Wang, Y.L.; Liang, T.G. Mapping snow depth distribution from 1980 to 2020 on the tibetan plateau using multi-source remote sensing data and downscaling techniques. ISPRS J. Photogramm. Remote Sens. 2023, 205, 246–262. [Google Scholar] [CrossRef]
  7. Luo, J.; Dong, C.; Lin, K.; Chen, X.; Zhao, L.; Menzel, L. Mapping snow cover in forests using optical remote sensing, machine learning and time-lapse photography. Remote Sens. Environ. 2022, 275, 113017. [Google Scholar] [CrossRef]
  8. Cannistra, A.F.; Shean, D.E.; Cristea, N.C. High-resolution CubeSat imagery and machine learning for detailed snow-covered area. Remote Sens. Environ. 2021, 258, 112399. [Google Scholar] [CrossRef]
  9. Du, J.; Watts, J.D.; Jiang, L.; Lu, H.; Cheng, X.; Duguay, C.; Tarolli, P. Remote sensing of environmental changes in cold regions: Methods, achievements and challenges. Remote Sens. 2019, 11, 1952. [Google Scholar] [CrossRef]
  10. Wu, X.; Shi, Z.; Zou, Z. A geographic information-driven method and a new large scale dataset for remote sensing cloud/snow detection. ISPRS J. Photogramm. Remote Sens. 2021, 174, 87–104. [Google Scholar] [CrossRef]
  11. Wu, X.; Zhu, R.; Long, Y.; Zhang, W. Spatial Trend and Impact of Snowmelt Rate in Spring across China’s Three Main Stable Snow Cover Regions over the Past 40 Years Based on Remote Sensing. Remote Sens. 2022, 14, 4176. [Google Scholar] [CrossRef]
  12. Paudel, K.P.; Andersen, P. Monitoring snow cover variability in an agropastoral area in the Trans Himalayan region of Nepal using MODIS data with improved cloud removal methodology. Remote Sens. Environ. 2011, 115, 1234–1246. [Google Scholar] [CrossRef]
  13. Crawford, C.J.; Manson, S.M.; Bauer, M.E.; Hall, D.K. Multitemporal snow cover mapping in mountainous terrain for Landsat climate data record development. Remote Sens. Environ. 2013, 135, 224–233. [Google Scholar] [CrossRef]
  14. Zakeri, F.; Mariethoz, G. Synthesizing long-term satellite imagery consistent with climate data: Application to daily snow cover. Remote Sens. Environ. 2024, 300, 113877. [Google Scholar] [CrossRef]
  15. Dong, J.; Walker, J.P.; Houser, P.R. Factors affecting remotely sensed snow water equivalent uncertainty. Remote Sens. Environ. 2005, 97, 68–82. [Google Scholar] [CrossRef]
  16. Jin, D.; Lee, K.S.; Choi, S.; Seong, N.H.; Jung, D.; Sim, S.; Woo, J.; Jeon, U.; Byeon, Y.; Han, K.S. An improvement of snow/cloud discrimination from machine learning using geostationary satellite data. Int. J. Digit. Earth 2022, 15, 2355–2375. [Google Scholar] [CrossRef]
  17. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015. In Proceedings of the 18th International Conference, Munich, Germany, 5–9 October 2015; Part III 18. Springer International Publishing: Cham, Switzerland, 2015. [Google Scholar]
  18. Chen, L.-C. Rethinking atrous convolution for semantic image segmentation. arXiv 2017, arXiv:1706.05587. [Google Scholar]
  19. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Advances in neural information processing systems. Adv. Neural Inf. Process. Syst. 2017, 30. [Google Scholar] [CrossRef]
  20. Wang, Y.; Su, J.; Zhai, X.; Meng, F.; Liu, C. Snow coverage mapping by learning from sentinel-2 satellite multispectral images via machine learning algorithms. Remote Sens. 2022, 14, 782. [Google Scholar] [CrossRef]
  21. Xing, D.; Hou, J.; Huang, C.; Zhang, W. Spatiotemporal reconstruction of MODIS normalized difference snow index products using U-Net with partial convolutions. Remote Sens. 2022, 14, 1795. [Google Scholar] [CrossRef]
  22. Yin, M.; Wang, P.; Ni, C.; Hao, W. Cloud and snow detection of remote sensing images based on improved Unet3+. Sci. Rep. 2022, 12, 14415. [Google Scholar] [CrossRef] [PubMed]
  23. Guo, X.; Chen, Y.; Liu, X.; Zhao, Y. Extraction of snow cover from high-resolution remote sensing imagery using deep learning on a small dataset. Remote Sens. Lett. 2020, 11, 66–75. [Google Scholar] [CrossRef]
  24. Kan, X.; Lu, Z.; Zhang, Y.; Zhu, L.; Sian, K.T.C.L.K.; Wang, J.; Liu, X.; Zhou, Z.; Cao, H. DSRSS-Net: Improved-Resolution Snow Cover Mapping from FY-4A Satellite Images Using the Dual-Branch Super-Resolution Semantic Segmentation Network. Remote Sens. 2023, 15, 4431. [Google Scholar] [CrossRef]
  25. Wang, Z.; Fan, B.; Tu, Z.; Li, H.; Chen, D. Cloud and Snow Identification Based on DeepLab V3+ and CRF Combined Model for GF-1 WFV Images. Remote Sens. 2022, 14, 4880. [Google Scholar] [CrossRef]
  26. Ding, L.; Xia, M.; Lin, H.; Hu, K. Multi-Level Attention Interactive Network for Cloud and Snow Detection Segmentation. Remote Sens. 2023, 16, 112. [Google Scholar] [CrossRef]
  27. Ma, J.; Shen, H.; Cai, Y.; Zhang, T.; Su, J.; Chen, W.H.; Li, J. UCTNet with dual-flow architecture: Snow coverage mapping with Sentinel-2 satellite imagery. Remote Sens. 2023, 15, 4213. [Google Scholar] [CrossRef]
  28. Zhang, G.; Gao, X.; Yang, Y.; Wang, M.; Ran, S. Controllably Deep Supervision and Multi-Scale Feature Fusion Network for Cloud and Snow Detection Based on Medium- and High-Resolution Imagery Dataset. Remote Sens. 2021, 13, 4805. [Google Scholar] [CrossRef]
  29. Hall, D.K.; Riggs, G.A.; Salomonson, V.V. Development of methods for mapping global snow cover using moderate resolution imaging spectroradiometer data. Remote Sens. Environ. 1995, 54, 127–140. [Google Scholar] [CrossRef]
  30. Chen, L.; Zhang, W.; Yi, Y.; Zhang, Z.; Chao, S. Long time-series glacier outlines in the three-rivers headwater region from 1986 to 2021 based on deep learning. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 5734–5752. [Google Scholar] [CrossRef]
  31. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  32. Zhang, X.D.; Zeng, H.; Zhang, L. Edge-oriented Convolution Block for Real-time Super Resolution on Mobile Devices. In Proceedings of the 29th ACM International Conference on Multimedia, Chengdu, China, 20–24 October 2021; Association for Computing Machinery: New York, NY, USA, 2021; pp. 4034–4043. [Google Scholar]
  33. Bjorck, N.; Gomes, C.P.; Selman, B.; Weinberger, K.Q. Understanding batch normalization. Adv. Neural Inf. Process. Syst. 2018, 31. [Google Scholar] [CrossRef]
  34. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  35. Chen, L.; Zhang, H.; Xiao, J.; Nie, L.; Shao, J.; Liu, W.; Chua, T.S. Sca-cnn: Spatial and channel-wise attention in convolutional networks for image captioning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, USA, 21–26 July 2017; pp. 5659–5667. [Google Scholar]
  36. Choudhary, V.; Guha, P.; Tripathi, K.; Mishra, S. Edge detection of variety of cowpea leaves using opencv and deep learning. In Proceedings of the 2022 4th International Conference on Advances in Computing, Communication Control and Networking (ICAC3N), Greater Noida, India, 16–17 December 2022; pp. 1312–1316. [Google Scholar]
  37. Rezatofighi, H.; Tsoi, N.; Gwak, J.; Sadeghian, A.; Reid, I.; Savarese, S. Generalized intersection over union: A metric and a loss for bounding box regression. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 658–666. [Google Scholar]
  38. Milletari, F.; Navab, N.; Ahmadi, S.A. V-net: Fully convolutional neural networks for volumetric medical image segmentation. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016; pp. 565–571. [Google Scholar]
  39. Badrinarayanan, V.; Kendall, A.; Cipolla, R. Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef] [PubMed]
  40. Zhan, Y.; Wang, J.; Shi, J.; Cheng, G.; Yao, L.; Sun, W. Distinguishing cloud and snow in satellite images via deep convolutional network. IEEE Geosci. Remote Sens. Lett. 2017, 14, 1785–1789. [Google Scholar] [CrossRef]
  41. Xia, M.; Liu, W.A.; Shi, B.; Weng, L.; Liu, J. Cloud/snow recognition for multispectral satellite imagery based on a multidimensional deep residual network. Int. J. Remote Sens. 2019, 40, 156–170. [Google Scholar] [CrossRef]
  42. Gupta, R.; Nanda, S.J. Cloud detection in satellite images with classical and deep neural network approach: A review. Multimed. Tools Appl. 2022, 22, 31847–31880. [Google Scholar] [CrossRef]
  43. Yao, X.; Guo, Q.; Li, A. Light-weight cloud detection network for optical remote sensing images with attention-based deeplabv3+ architecture. Remote Sens. 2021, 13, 3617. [Google Scholar] [CrossRef]
  44. Yang, J.; Guo, J.; Yue, H.; Liu, Z.; Hu, H.; Li, K. CDnet: CNN-based cloud detection for remote sensing imagery. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6195–6211. [Google Scholar] [CrossRef]
  45. Zhang, C.; Bengio, S.; Hardt, M.; Recht, B.; Vinyals, O. Understanding deep learning (still) requires rethinking generalization. Commun. ACM 2021, 64, 107–115. [Google Scholar] [CrossRef]
  46. Wang, Y.; Gu, L.; Li, X.; Gao, F.; Jiang, T. Coexisting cloud and snow detection based on a hybrid features network applied to remote sensing images. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5405515. [Google Scholar] [CrossRef]
  47. Salomonson, V.V.; Appel, I. Estimating fractional snow cover from MODIS using the normalized difference snow index. Remote Sens. Environ. 2004, 89, 351–360. [Google Scholar] [CrossRef]
  48. Torralbo, P.; Pimentel, R.; Polo, M.J.; Notarnicola, C. Characterizing Snow Dynamics in Semi-Arid Mountain Regions with Multitemporal Sentinel-1 Imagery: A Case Study in the Sierra Nevada, Spain. Remote Sens. 2023, 15, 5365. [Google Scholar] [CrossRef]
Figure 1. True color CSWV_S6 data synthesized from the red, green, and blue bands (the numbering in the figure corresponds to the original naming in the acquired files).
Figure 1. True color CSWV_S6 data synthesized from the red, green, and blue bands (the numbering in the figure corresponds to the original naming in the acquired files).
Remotesensing 16 03430 g001
Figure 2. (a) RGB composite of Landsat 8 imagery (red: band 4, green: band 3, blue: band 2). (b) Land cover types.
Figure 2. (a) RGB composite of Landsat 8 imagery (red: band 4, green: band 3, blue: band 2). (b) Land cover types.
Remotesensing 16 03430 g002
Figure 3. CEFCSAU-net network model architecture. The input size was (512, 512, C), where C denotes the number of channels, and experiments in this paper utilized either 3 or 4; during the model’s operation on a GPU, intermediate feature maps were stored as tensors.
Figure 3. CEFCSAU-net network model architecture. The input size was (512, 512, C), where C denotes the number of channels, and experiments in this paper utilized either 3 or 4; during the model’s operation on a GPU, intermediate feature maps were stored as tensors.
Remotesensing 16 03430 g003
Figure 4. Attention mechanism module for channel and space mixing. Here, (H, W, C) represent the height, width, and number of channels of the feature data, respectively, with values determined by input features at different stages. CF and CF’ denote feature maps from various intermediate operations within the channel attention mechanism. Cat Sf, Sf, Sf’ represent feature maps from different intermediate operations of the spatial attention mechanism. The SA feature denotes the feature map post-spatial attention mechanism, the CA feature represents those post-channel attention mechanisms, and the CSA feature illustrates feature maps following the CSA module.
Figure 4. Attention mechanism module for channel and space mixing. Here, (H, W, C) represent the height, width, and number of channels of the feature data, respectively, with values determined by input features at different stages. CF and CF’ denote feature maps from various intermediate operations within the channel attention mechanism. Cat Sf, Sf, Sf’ represent feature maps from different intermediate operations of the spatial attention mechanism. The SA feature denotes the feature map post-spatial attention mechanism, the CA feature represents those post-channel attention mechanisms, and the CSA feature illustrates feature maps following the CSA module.
Remotesensing 16 03430 g004
Figure 5. Cross-scale edge-aware feature fusion module. Sobelx F, Sobely F, and Laplacian F denote feature maps resulting from various edge detection operations. Shallow F refers to feature maps following shallow feature convolution. Fusion F illustrates feature maps resulting from the fusion of shallow and deep features. Deep F’ represents feature maps after a series of operations on deep features.
Figure 5. Cross-scale edge-aware feature fusion module. Sobelx F, Sobely F, and Laplacian F denote feature maps resulting from various edge detection operations. Shallow F refers to feature maps following shallow feature convolution. Fusion F illustrates feature maps resulting from the fusion of shallow and deep features. Deep F’ represents feature maps after a series of operations on deep features.
Remotesensing 16 03430 g005
Figure 6. Snow extraction results of CSWV_S6 data on different segmentation models (a set of two rows is the same image data, and rows two, four, and six are zoomed-in images of local details corresponding to rows one, three, and five. The blue area is snow, the white area is non-snow, and the red area is false detection).
Figure 6. Snow extraction results of CSWV_S6 data on different segmentation models (a set of two rows is the same image data, and rows two, four, and six are zoomed-in images of local details corresponding to rows one, three, and five. The blue area is snow, the white area is non-snow, and the red area is false detection).
Remotesensing 16 03430 g006
Figure 7. Snow extraction results from different deep learning models for Landsat 8 OLI imagery (blue areas are snow, white areas are non-snow, and red areas are false detections).
Figure 7. Snow extraction results from different deep learning models for Landsat 8 OLI imagery (blue areas are snow, white areas are non-snow, and red areas are false detections).
Remotesensing 16 03430 g007
Figure 8. (a) CSWV_S6 test set scores on different models for each type of metrics and (b) Landsat 8 OLI test set scores for various metrics on different models.
Figure 8. (a) CSWV_S6 test set scores on different models for each type of metrics and (b) Landsat 8 OLI test set scores for various metrics on different models.
Remotesensing 16 03430 g008
Figure 9. Score results of the three CSWV_S6 test sets’ example data on the evaluation metrics on each model, with 0.08% of snow image elements in the first row of data, 0.95% of snow image elements in the second row of data, and 1.73% of data in the third row of data.
Figure 9. Score results of the three CSWV_S6 test sets’ example data on the evaluation metrics on each model, with 0.08% of snow image elements in the first row of data, 0.95% of snow image elements in the second row of data, and 1.73% of data in the third row of data.
Remotesensing 16 03430 g009
Figure 10. Map of CEFCSAU-net model’s snow extraction in the cloud–snow confusion scenario of CSWV_S6 test set.
Figure 10. Map of CEFCSAU-net model’s snow extraction in the cloud–snow confusion scenario of CSWV_S6 test set.
Remotesensing 16 03430 g010
Figure 11. Heat map comparing the mean values of ablation experiments on the test set with different data sets: (a) CSWV_6 dataset, (b) Landsat8 OLI dataset.
Figure 11. Heat map comparing the mean values of ablation experiments on the test set with different data sets: (a) CSWV_6 dataset, (b) Landsat8 OLI dataset.
Remotesensing 16 03430 g011
Figure 12. (a) Input data image; (b) feature map of the first 8 channels before the intermediate feature data first pass through the CSA module; (c) feature map of the first 8 channels after the feature data first pass through the CSA module.
Figure 12. (a) Input data image; (b) feature map of the first 8 channels before the intermediate feature data first pass through the CSA module; (c) feature map of the first 8 channels after the feature data first pass through the CSA module.
Remotesensing 16 03430 g012
Figure 13. This figure displays average feature maps following skip connections at various stages under different configurations of the CEFCSAU-net model. In this figure, (a1a4) represent the model configuration without both the CSA and CEF modules; (b1b4) indicate configurations without the CSA module yet including the CEF module; and (c1c4) depict configurations featuring both CSA and CEF modules. The dimensions of the four columns of feature maps are sequentially 512 × 512, 256 × 256, 128 × 128, and 64 × 64.
Figure 13. This figure displays average feature maps following skip connections at various stages under different configurations of the CEFCSAU-net model. In this figure, (a1a4) represent the model configuration without both the CSA and CEF modules; (b1b4) indicate configurations without the CSA module yet including the CEF module; and (c1c4) depict configurations featuring both CSA and CEF modules. The dimensions of the four columns of feature maps are sequentially 512 × 512, 256 × 256, 128 × 128, and 64 × 64.
Remotesensing 16 03430 g013
Table 1. Comparison of mean values of different segmentation models on CSWV_S6 test set data.
Table 1. Comparison of mean values of different segmentation models on CSWV_S6 test set data.
ModelPA (%)R (%)F1 (%)mIoU (%)
FCN8s96.3171.8669.9675.69
SegNet94.9979.4276.978.56
U-net96.3472.2974.5479.03
DeepLab V3+97.5182.3881.40 83.77
CEFCSAU-net98.1487.8785.6286.28
Note: The bold text in the table represents the maximum value of the indicator in that column.
Table 2. Comparison of mean values of evaluation metrics on different segmentation models for Landsat test set data.
Table 2. Comparison of mean values of evaluation metrics on different segmentation models for Landsat test set data.
ModelPA (%)R (%)F1 (%)mIoU (%)
FCN8s85.6580.9583.6268.24
SegNet92.3691.9992.4282.09
U-net91.8993.6891.8981.26
DeepLab V3+92.6892.9192.3981.79
CEFCSAU-net95.6796.6895.7488.68
Note: The bold text in the table represents the maximum value of the indicator in that column.
Table 3. Comparison of the mean values of different segmentation model evaluation metrics on the test set after the introduction of SWIR 1 band to Landsat data.
Table 3. Comparison of the mean values of different segmentation model evaluation metrics on the test set after the introduction of SWIR 1 band to Landsat data.
ModelPA (%)R (%)F1 (%)mIoU (%)
FCN8s84.9686.4883.8665.22
SegNet94.1291.4393.9885.68
U-net91.6391.9191.8781.11
DeepLab V3+92.1894.6492.0880.38
CEFCSAU-net96.6597.9796.7590.71
Note: The bold text in the table represents the maximum value of the indicator in that column.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yu, Z.; Gong, H.; Zhang, S.; Wang, W. Snow Cover Extraction from Landsat 8 OLI Based on Deep Learning with Cross-Scale Edge-Aware and Attention Mechanism. Remote Sens. 2024, 16, 3430. https://doi.org/10.3390/rs16183430

AMA Style

Yu Z, Gong H, Zhang S, Wang W. Snow Cover Extraction from Landsat 8 OLI Based on Deep Learning with Cross-Scale Edge-Aware and Attention Mechanism. Remote Sensing. 2024; 16(18):3430. https://doi.org/10.3390/rs16183430

Chicago/Turabian Style

Yu, Zehao, Hanying Gong, Shiqiang Zhang, and Wei Wang. 2024. "Snow Cover Extraction from Landsat 8 OLI Based on Deep Learning with Cross-Scale Edge-Aware and Attention Mechanism" Remote Sensing 16, no. 18: 3430. https://doi.org/10.3390/rs16183430

APA Style

Yu, Z., Gong, H., Zhang, S., & Wang, W. (2024). Snow Cover Extraction from Landsat 8 OLI Based on Deep Learning with Cross-Scale Edge-Aware and Attention Mechanism. Remote Sensing, 16(18), 3430. https://doi.org/10.3390/rs16183430

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop