Next Article in Journal
InSAR-Based Multi-Source Monitoring and Modeling of Multi-Seam Mining-Induced Deformation and Hazard Chain Evolution in the Loess Gully Region
Previous Article in Journal
Spectral Unmixing of Coastal Dune Plant Species from Very High Resolution Satellite Imagery
Previous Article in Special Issue
Improvement of Snow Albedo Simulation Considering Water Content
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optimizing Cloud Mask Accuracy over Snow-Covered Terrain with a Multistage Decision Tree Framework

1
Key Laboratory of Cryospheric Science and Frozen Soil Engineering, Northwest Institute of Eco-Environment and Resources, Chinese Academy of Sciences, Lanzhou 730000, China
2
State Key Laboratory of Geographic Information Science and Technology, Institute of Geographic Sciences and Natural Resources Research, Chinese Academy of Sciences, Beijing 100101, China
3
Heihe Remote Sensing Experimental Research Station, Northwest Institute of Eco-Environment and Resources, Chinese Academy of Sciences, Lanzhou 730000, China
4
University of Chinese Academy of Sciences, Beijing 100049, China
5
College of Earth and Environmental Sciences, Lanzhou University, Lanzhou 730000, China
6
Meteorological Institute of Qinghai Province, Xining 810001, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(24), 3992; https://doi.org/10.3390/rs17243992
Submission received: 25 September 2025 / Revised: 4 December 2025 / Accepted: 6 December 2025 / Published: 10 December 2025
(This article belongs to the Special Issue Remote Sensing Modelling and Measuring Snow Cover and Snow Albedo)

Highlights

What are the main findings?
  • Optimized a multi-level decision tree cloud detection algorithm for cloud-snow discrimination.
  • Utilized the distinct bright temperature difference at 3.7 μm and 11 μm to enhance cloud detection performance.
What are the implications of the main findings?
  • Outperforms existing algorithms and significantly reduces the false cloud detection in snow-covered areas.
  • Provides accurate and efficient cloud detection in the Northern Hemisphere to support related cryospheric studies.

Abstract

High-resolution optical remote sensing imagery plays a crucial role in monitoring the Earth’s surface. However, cloud obstruction and spectral confusion between clouds and snow significantly compromise data quality and application reliability, leading to persistent cloud overestimation in optical remote sensing products. To address this challenge, this study developed an enhanced multi-threshold cloud detection algorithm based on AVHRR surface reflectance data, which incorporates dynamic threshold optimization within a multi-level decision tree framework. Utilizing Landsat 5 SR as reference data, the algorithm demonstrated superior cloud-snow discrimination capability, achieving an overall accuracy (OA) of 82.08%, with the user’s accuracy (UA) and F-score reaching 79.41% and 82.55%. Comparative evaluation demonstrates that the proposed algorithm outperforms two existing algorithms, with OA improvements of 17.42% and 7.93%, respectively. A particularly notable enhancement is the significant reduction in cloud misidentification, as reflected by UA increases of 21.02% and 13.21%. These improvements are most pronounced in high-altitude mountainous regions with snow cover. The algorithm maintains computational efficiency while providing reliable cloud masking, thereby offering enhanced support for snow cover monitoring and broader environmental applications.

1. Introduction

High-resolution optical remote sensing imagery has become an indispensable data source for monitoring dynamic changes on the Earth’s surface, widely applied in climate research, hydrological modeling, and environmental change analysis [1,2]. However, due to atmospheric conditions and sensor performance limitations, these images often suffer from data loss and quality deterioration caused by phenomena such as cloud cover, cloud shadows, and polar night [3,4]. This issue is particularly pronounced in regions with seasonally accumulated snow, where clouds and snow exhibit high spectral similarity in the visible to near-infrared bands, complicating the differentiation between the two [5]. Inaccurate cloud detection can lead to effective pixels being incorrectly obscured, reducing data availability, and may also cause overestimation of surface reflectance due to residual sub-pixel clouds or cloud shadows, thereby distorting time series analysis results and posing a significant threat to the accuracy of key geographical parameters, such as snow cover area [6].
To achieve effective cloud masking, various algorithms have been developed. Among them, threshold-based spectral methods distinguish clouds from the surface by setting reflectance thresholds. Although straightforward to implement, their accuracy is limited under complex atmospheric conditions, such as in mixed-pixel scenarios [7,8]. Other approaches, including the CMAPROB algorithm based on naive Bayesian theory [9] and optimal estimation (OE) techniques [10], offer more flexible identification strategies that can effectively handle different cloud types. However, these methods require higher data quality and computational resources. In recent years, intelligent processing techniques such as machine learning have improved recognition accuracy to some extent by training models to learn cloud features, while also providing long-term and comprehensive cloud type distribution information [11]. Nevertheless, class imbalance in datasets—where certain cloud types occur less frequently—can lead to model bias toward more common types. Additionally, inconsistencies among satellite products may affect the reliability of the results.
For large-scale surface parameter retrieval in the Northern Hemisphere, the long-term surface reflectance data acquired by the Advanced Very High-Resolution Radiometer (AVHRR) sensor are of irreplaceable value [12,13]. With over forty years of temporal coverage and near-global spatial availability, these data provide a critical foundation for monitoring land surface dynamics such as vegetation change, land cover classification, and snow/ice monitoring [14,15]. A cloud identification algorithm and threshold system based on AVHRR radiometric calibration data was proposed by Hori et al. [16] and has been widely applied in Northern Hemisphere snow product generation. Although this algorithm employed conventional cloud detection tests with enhanced strictness, it still inadequately addressed the spectral confusion between clouds and snow, resulting in both misclassification and omission errors. In response, Hao et al. [17] optimized the cloud masking thresholds within Hori’s framework using the National Oceanic and Atmospheric Administration (NOAA) Climate Data Record (CDR) of AVHRR Surface Reflectance Version 4 (AVHRR SR V4) over China, achieving an overall accuracy (OA) exceeding 84.76% and demonstrating the value of region-specific adaptive optimization. However, since the algorithm was developed and validated exclusively over China using AVHRR SR V4, its applicability to other geographical regions, diverse surface types, and subsequent CDR versions has not been verified, which substantially limits its utility for global-scale parameter retrieval.
The recently released NOAA CDR of AVHRR Surface Reflectance Version 5 (AVHRR SR V5) [18], jointly developed by NASA’s Goddard Space Flight Center (GSFC) and the University of Maryland (UMD), incorporates significant improvements in radiometric calibration and atmospheric correction. However, studies have indicated that the built-in cloud detection in this version exhibits biases, particularly a notable overestimation of cloud coverage. As shown in Figure 1 (taking 18 November 1988, as an example), obvious over-identification of clouds is observed in the AVHRR SR V5 cloud mask. Such biases can severely impact the quality of various land surface parameter retrieval products. The data exhibit pronounced band fluctuations before and after 2000, while the period from 1981 to 2000 demonstrates relatively minor fluctuations and greater stability. Given this early stability, alongside the availability of higher-quality satellite data from sensors such as MODIS, VIIRS, and Sentinel after 2000, and considering that AVHRR was among the earliest satellite systems to offer extensive spatial coverage with relatively high spatiotemporal resolution, there is an urgent need to develop a high-accuracy cloud detection algorithm adapted to the characteristics of AVHRR SR V5 with applicability in the Northern Hemisphere from 1981 to 2000, to enhance the reliability and consistency of long-term land surface parameter products.
This study proposes an improved cloud detection algorithm based on multi-source data fusion to address misidentification issues caused by cloud-snow spectral confusion in northern regions. Using Landsat high-resolution imagery as “truth” reference, combined with an improved SNOMAP algorithm [19] and cloud mask information from the QA_PIXEL quality band [20], a cloud-snow sample dataset integrating spectral features and topographic parameters was constructed. By analyzing the spectral response characteristics of clouds and snow from visible to shortwave infrared bands, a multi-level decision tree algorithm was established through threshold optimization strategies to accurately identify cloud pixels in the Northern Hemisphere from 1981 to 2000. Finally, the performance of the proposed algorithm was evaluated, and its uncertainties as well as future application prospects were analyzed.

2. Materials and Preprocessing

2.1. AVHRR Surface Reflectance CDR

The National Oceanic and Atmospheric Administration (NOAA) Climate Data Record (CDR) of Advanced Very High-resolution Radiometer (AVHRR) Surface Reflectance Version 5 (AVHRR SR V5) images serve as the foundational dataset for this study. Produced by the NASA Goddard Space Flight Center (Greenbelt, MD, USA) and the University of Maryland (College Park, MD, USA), this dataset is archived at the NOAA National Centers for Environmental Information (Keesler Air Force Base, MS, USA) and encompasses gridded daily surface reflectance, brightness temperatures, and quality control flags (QA) from 24 June 1981 to 31 December 2000. The data, derived from AVHRR sensors onboard eight NOAA polar-orbiting satellites, was generated using AVHRR GAC Level 1B data through geolocation, calibration, and atmospheric correction, offering a global coverage at a spatial resolution of 0.05° [18]. Compared to Version 4, Version 5 incorporates significant scientific enhancements, including higher resolution ancillary data and improved methodologies for bidirectional reflectance distribution function (BRDF) correction, calibration, compositing, and QA. Additionally, it rectifies known errors in temporal, latitudinal, and longitudinal variables while enhancing global and variable attribute definitions. This dataset provides essential model input, including surface reflectance data, brightness temperature data, angle data, and quality control information, with spectral band details outlined in Table 1. The quality control flags are summarized in Table 2. The processing of all AVHRR SR V5 images was conducted using the Google Earth Engine (GEE) cloud platform (Google Inc., Mountain View, CA, USA) [21].
Cloudy pixels are identified by comparing the spectrally adjusted AVHRR data—adjusted by a factor of 0.97—to a MODIS monthly average red band (MODIS Channel 1) BRDF corrected climatology, computed using a decade of MODIS data. A pixel is classified as cloudy if the difference exceeds 0.03. Furthermore, cloud height is estimated based on temperatures derived from AVHRR Channels 4 and 5, while shadow pixels are determined by the projection of cloudy pixels onto the surface, retaining only the cloudy flag for those identified as both cloudy and shadowed [20].

2.2. Landsat 5 Reference Snow and Cloud Maps

The Landsat data utilized in this study were primarily obtained from the Landsat 5 Level 2, Collection 2, Tier 1 Surface Reflectance (SR) products distributed by the United States Geological Survey (Reston, VA, USA) [22]. This dataset is generated from the Thematic Mapper (TM) sensor onboard the Landsat 5 satellite and includes atmospherically corrected surface reflectance and land surface temperature data. The Landsat 5 satellite, the fifth satellite in the Landsat program, was launched from Vandenberg Air Force Base in California on 1 March 1984 and was decommissioned on 5 June 2013. The image data comprises four visible and near-infrared (VNIR) bands and two shortwave infrared (SWIR) bands, all processed to generate orthorectified surface reflectance, along with one thermal infrared (TIR) band processed for orthorectified surface temperature. Detailed band information is presented in Table 3. The dataset also includes intermediate bands and quality control (QA) bands utilized for generating land surface temperature products, alongside pixel quality attributes derived from the CFMASK algorithm [23] for cloud information extraction, as outlined in Table 4. The Landsat 5 Surface Reflectance (SR) products are generated using the Landsat Ecosystem Disturbance Adaptive Processing System (LEDAPS) algorithm (version 3.4.0), while the Collection 2 Surface Temperature (ST) products are derived from a single-channel algorithm jointly developed by the Rochester Institute of Technology (RIT, Rochester, NY, USA) and the Jet Propulsion Laboratory (JPL, Pasadena, CA, USA) of NASA. Data are organized through a standardized reference grid and packaged into overlapping “scenes” covering approximately 170 km × 183 km.
Landsat imagery is widely regarded as a “true” reference dataset for medium and low-resolution remote sensing data due to its rich spectral information, high spatial resolution, long time series, and free accessibility [24,25]. In this study, snow and cloud samples within the multi-level decision tree cloud detection algorithm reference the Landsat 5 imagery as ground truth.
To generate the reference snow cover data, an improved “SNOMAP” algorithm was employed. This method addresses the limitations of the traditional NDSI-based approach (NDSI ≥ 0.4 and NIR reflectance ≥ 0.11), which often underestimates snow cover in forested areas due to canopy obstruction. The improved algorithm incorporates multiple criteria, specifically NDFSI > 0.4 and NDVI < 0.6, to effectively identify snow pixels in forests where NDSI values fall below 0.4, resulting in a 30 m binary snow cover map [19].
Cloud reference data were obtained from the “QA_PIXEL” quality assessment band of Landsat 5 imagery, which provides reliable cloud masking information. To match the spatial resolution of AVHRR products, both the binary snow maps and cloud masks were aggregated to a 5 km resolution. Each 5 km pixel contains a value indicating either snow or no snow, and cloud or no cloud, serving as reference truth data for subsequent analysis.
Figure 2 illustrates the generation process of the reference snow data. Using Landsat 5 imagery (path/row 045024) from 30 April 1998, as an example, it displays the 30 m false-color composite, the binary snow cover map derived from the improved algorithm, and the aggregated 5 km resolution reference map. Similarly, Figure 3 presents the 30 m false-color image alongside its corresponding cloud mask derived from the “QA_PIXEL” band, as well as the aggregated 5 km cloud reference data. These aggregated datasets provide essential ground truth for the development and validation of cloud and snow discrimination models.

2.3. Ancillary Data

The GTOPO30 dataset is a global digital elevation model (DEM) developed by the U.S. Geological Survey (USGS, Reston, VA, USA), offering a horizontal resolution of 30 arc-seconds (approximately 1 km) and a vertical accuracy of ±30 m [26]. It provides essential elevation data as input to the cloud detection model. To align with the spatial resolution of AVHRR products, the GTOPO30 data were resampled to 5 km using bicubic interpolation.
The MOD44W Version 6 product provides a global land-water mask at 250 m resolution, generated through a decision-tree classification applied to MODIS data and validated against its predecessor, MOD44W V5. The classification incorporates multiple ancillary masks to mitigate errors caused by terrain shadows, burn scars, clouds, and oceanic ice cover. This mask was resampled to 5 km using nearest-neighbor interpolation to maintain consistency with the AVHRR grid [27].

3. Methods

A multi-layer decision tree-based framework was designed to enhance cloud masking accuracy over snow-covered terrain, which consists of four main steps (Figure 4).
(1) Selection of reference truth samples. The selection of reference truth samples was conducted on the Google Earth Engine (GEE) platform using Landsat 5 imagery to acquire high-confidence reference data for clouds and snow. Based on predefined criteria, a large set of reference truth labels was generated to provide a reliable truth reference for subsequent model training and validation.
(2) Build training dataset. Building on the existing AVHRR cloud discrimination algorithm framework, spectral and topographic parameters were derived from the reference truth data. Feature screening was then applied to yield a representative and quality-controlled training dataset as input parameters for the decision tree algorithm construction.
(3) Algorithm improvement based on threshold optimization. The cloud detection algorithm was enhanced through the application of threshold optimization strategies. Thresholds for ten cloud detection schemes were specifically improved, focusing on the key feature BT3-BT4 for distinguishing between cloud and snow. This led to the development of a high-precision cloud recognition algorithm capable of effectively differentiating cloud pixels from snow.
(4) Accuracy Assessment. Model performance was quantitatively evaluated using accuracy assessment metrics, and potential sources of uncertainty in cloud-related studies were analyzed. Optimization suggestions were provided for the algorithms and models developed in this chapter to enhance their reliability and stability under complex scenarios. The performance was further assessed by examining algorithmic uncertainties and exploring future application prospects.

3.1. Selection of Truth Reference Samples

High-resolution Landsat 5 SR product serves as a “true” reference for mid-resolution data in the classification of clouds and snow in this study. The specific methods for obtaining “true” maps for cloud and snow can be found in Section 2.2. To select the most representative cloud and snow sample images, the following screening criteria were applied: the period from 1 June to 1 August was designated as the training selection date for cloud samples, while 1 October to 1 May of the following year was allocated for snow sample selection. This choice is primarily due to the generally stable climatic conditions during the June to August period, which favor consistent cloud formation and distribution patterns. Additionally, the infrequent snow events during this timeframe help reduce confusion between clouds and snow, enhancing the representativeness and effectiveness of the samples. Conversely, the period from 1 October to 1 May is characterized by lower temperatures and frequent snowfall, creating favorable natural conditions for the formation and accumulation of snow.
Based on these criteria, a total of 17,851 cloud samples and 5992 snow samples were screened. During the sample selection process, relatively lenient standards were adopted to ensure diversity and quantity in the samples. This approach aims to improve the cloud detection algorithm through the use of extensive sample data. The advantage of this strategy lies in its ability to enhance the generalization capability of the algorithm by encompassing a wide range of climatic and environmental conditions, thereby ensuring more stable performance in practical applications. Furthermore, a diverse dataset can effectively improve the algorithm’s ability to distinguish between clouds and snow against complex backgrounds.

3.2. Build Training Dataset

To enhance the accuracy and reliability of cloud detection algorithms, a new algorithm was developed based on traditional combinations of existing methods [7,8,16,28,29]. During the construction process, surface reflectance (SR), brightness temperature (BT), Index and band combination data (I&B), and terrain data (TD) were utilized, comprising nine key variables: SR1, SR2, SR3, BT4, the brightness temperature difference between BT3 and BT4 (BT3-BT4), the brightness temperature difference between BT4 and BT5 (BT4-BT5), the reflectance difference between SR1 and SR2 (SR1-SR2), the Normalized Difference Vegetation Index (NDVI), and elevation (DEM), as summarized in Table 5.
The selection of these variables aims to comprehensively reflect the spectral and thermal characteristics of the cloud layer and the underlying surface. Surface reflectance data (SR1, SR2, SR3) provide spectral information across different bands, aiding in the identification of cloud types and their distribution [30]. The BT series indicators offer critical data support for the thermal characteristics of clouds, particularly the brightness temperature difference between BT3 and BT4, which serves as a primary testing metric to effectively distinguish clouds from other surface covers [31]. The reflectance difference (Ref1-Ref2) and brightness temperature difference (BT4-BT5) capture subtle variations between cloudy and cloud-free areas. Additionally, the NDVI and elevation (DEM) provide vegetation cover information and terrain characteristics, further enhancing the discriminative capacity of cloud-snow differentiation [32].
To enhance sample representativeness, invalid pixels were classified as gap pixels during the preprocessing stage and subsequently removed. These gap pixels included (1) pixels with poor quality control, (2) pixels with surface reflectance values outside the range [0, 1], (3) pixels exhibiting anomalous brightness temperatures, (4) water bodies, and (5) pixels with no data.

3.3. Algorithm Improvement Based on Threshold Optimization

3.3.1. Existing Algorithm Framework

Currently, the cloud detection algorithm based on AVHRR data was proposed by Hori et al. [16] and employs a multi-level threshold classification method to effectively identify cloud pixels. The core of this algorithm lies in the hierarchical classification of cloud pixels through stringent threshold settings, thereby enhancing the accuracy and reliability of the cloud detection process.
Initially, a preliminary classification is conducted based on specific thresholds of elevation DEM and the brightness temperature band BT4 at a wavelength of 11 μm: when the DEM value exceeds 300 m and the BT4 is below 260 K, the cloud pixels are categorized as the high or cold land (Target A); otherwise, they are classified as the plains or normal-temperature land (Target B).
Building on the preliminary classification, a secondary classification is performed using multiple variables, including SR1, SR2, SR3, BT4, BT4-BT5, SR1-SR2, and NDVI. In this process, clouds in high-altitude cold regions are further subdivided into four subclasses (A1, A2, A3, A4), while flat temperate regions are divided into six subclasses (B1, B2, B3, B4, B5, B6). Among these classifications, the brightness temperature difference between BT3 and BT4 (BT3-BT4) serves as a crucial third-level discriminative indicator, effectively distinguishing between different types of clouds. The specific framework of the algorithm is illustrated in Figure 5.
Despite the thresholds employed by this cloud detection algorithm being largely based on traditional cloud detection test combinations from previous studies, certain limitations persist. In particular, the setting of the BT3-BT4 threshold, while utilizing a more stringent cloud recognition algorithm, has not sufficiently addressed the confusion between clouds and snow, leading to potential misclassification and omission in practical applications. Therefore, optimization of the algorithm is urgently needed to enhance its accuracy and reliability under different environmental conditions.
In this context, the research team has built upon the aforementioned framework, utilizing AVHRR CDR SR V4 data to improve and evaluate the cloud detection algorithm thresholds specifically for the Chinese region [17]. Results demonstrate that the modified algorithm achieves an overall accuracy (OA) of 84.76% for cloud detection in the study area, further validating the effectiveness of the framework. However, it is important to note that the construction and application of this algorithm are primarily focused on the Chinese region and are based on AVHRR CDR SR V4, which somewhat limits its applicability in other geographical areas or with different versions of AVHRR CDR products.

3.3.2. Proposed Algorithm Improvement Strategy

Based on the AVHRR CDR SR V5 dataset and utilizing cloud and snow samples extracted from Landsat 5 SR maps as reference truth values, an improved threshold-optimized cloud detection algorithm was developed within the framework established by Hori et al. The study area was divided into two distinct regions: high-altitude mountainous areas (Target A) and temperate plains (Target B). Target A encompasses four cloud detection scenarios (A1–A4), while Target B includes six cloud detection scenarios (B1–B6) along with four clear-sky detection scenarios.
To address cloud-snow confusion over complex terrain, a third-level decision rule based on the BT3-BT4 threshold was incorporated into the original framework, specifically for cloud type A4 in high-altitude regions and type B6 in temperate plains. The multi-level decision tree cloud detection algorithm was implemented as follows: after data preprocessing, first- and second-level features were used to classify valid pixels into the ten cloud detection schemes (A1–A4 and B1–B6). Subsequently, frequency distributions of BT3-BT4 under different cloud categories were derived from cloud and snow training samples. Based on the distribution patterns of cloud and snow samples in the AVHRR SR V5 data, the overall accuracy (OA) was computed at intervals of 0.01, and the threshold corresponding to the maximum OA was selected as optimal.
Figure 4 illustrates the optimal determination scheme for the brightness temperature difference (BT3-BT4), using Target A1 as an example. The left panels show the frequency distributions of BT3-BT4BT3-BT4 for cloud and snow training samples from before the year 2000. The right panels display the variation in overall accuracy across different BT3-BT4 thresholds. For cloud type A1, the highest accuracy (84.81%) was achieved at the intersection of the frequency distributions of snow and cloud, corresponding to a BT3-BT4 threshold of 20 K. This value reflects a trade-off between the omission error (9.43%) and the commission error (19.08%). Thus, the final threshold was set to 20 K, whereby pixels with BT3-BT4 > 20 K are classified as cloud. Optimal thresholds for other cloud types were determined using the same methodology.

3.4. Accuracy Assessment Metrics

To comprehensively analyze the accuracy of classification results and the types of errors, this study employs a confusion matrix as a fundamental tool. The confusion matrix compares the predicted values with the actual values to provide a detailed display of the model’s misclassification across different categories. The specific form of the confusion matrix is shown in Table 6.
In this matrix, TP represents the number of samples that are actually positive and predicted as positive, while FN indicates the number of actual positives predicted as negatives; FP denotes the number of actual negatives predicted as positives, and TN represents the number of actual negatives predicted as negatives. In the multi-level decision tree cloud detection algorithm, the positive class refers to the identified clouds. Based on the confusion matrix, this study employs overall accuracy (OA), producer’s accuracy (PA), user’s accuracy (UA), and F-score (FS) metrics for accuracy assessment. OA indicates the proportion of correctly predicted values among all predictions, PA reflects the proportion of correctly predicted positives among actual positive cases, UA signifies the proportion of actual positives among predicted positives, and FS considers both UA and PA to evaluate the overall performance of the binary classification model. The specific calculation formulas are as follows:
OA = TP + TN TP + TN + FP + FN
PA = TP TP + FN
UA = TP TP + FP
FS = 2 × UA × PA UA + PA

4. Results

4.1. Improved Thresholds

Table 7 presents the improved cloud detection scheme for the Northern Hemisphere based on AVHRR CDR SR V5 data, which includes 10 cloud detection schemes and 4 clear sky schemes. Target A represents the high or cold land, comprising four types: A1–A4. Target B indicates the plains or normal-temperature land (other land), including ten cloud detection schemes: B1–B10. For each target, cloud detection tests are conducted from the top of the list to the bottom. If the cloud flag switch is set to “on” pixels are classified as cloudy when the threshold tests meet the conditions listed on the right. If the switch is set to “off” pixels previously identified as cloudy in earlier tests are reset to clear.

4.2. Accuracy Assessment

4.2.1. Overview of the Algorithm Accuracies

The average OA of the algorithm is 82.08%, with a PA of 86.35%, a UA of 79.41%, and an FS of 82.55%. These results, indicate that while the enhanced cloud detection algorithm still experiences some misclassifications and omissions in distinguishing between clouds and snow, it achieves a relatively high accuracy overall. The algorithm maintains a good balance between precision and recall, as evidenced by the F-Score. This reflects the effectiveness of the threshold optimization strategy applied to the BT3-BT4 spectral features, significantly improving the distinguishability between cloud and snow pixels.

4.2.2. Algorithm Accuracies of Different Cloud Detection Schemes

Figure 6 presents the accuracy evaluation results for ten cloud detection schemes, revealing a significant difference in overall accuracy between the high or cold land (Target A) and the plains or normal-temperature land (Target B). The target A1 achieves an OA of 84.81% and a PA of 90.57%, indicating strong cloud recognition capabilities; however, the UA of 80.92% suggests some misclassification issues. The OA for target A2 decreases slightly to 79.44%, with an increase in both omission and commission errors, reflecting the complexity of cloud-snow confusion in high-altitude regions. The target A3 exhibits a significantly lower OA of 67.43%, demonstrating insufficient recognition ability, with high rates of both omission and commission errors, particularly the latter. Although the target A4 shows some improvement over A3, further enhancements are still needed.
In contrast, the cloud detection schemes in Target B generally perform better. The target B1 achieves an OA of 83.27%, demonstrating good performance in avoiding omissions, although the commission rate remains relatively high. The target B2 reports an OA of 80.50%, slightly lower than target B1, while the target B3 reaches an OA of 83.64%, showing a favorable balance between omission and commission errors. The OAs for Target B4 and Target B5 are 87.16% and 86.09%, respectively, both reflecting high cloud recognition capabilities. Notably, the target B6 achieves an OA of 90.93%, with minimal omission errors and relatively few commission errors, highlighting its superiority in cloud detection.
The accuracy of Target B is significantly better than that of Target A, which can primarily be attributed to environmental factors. Target A faces complex climatic conditions and variable topographic features. For instance, the severe climatic fluctuations in high-altitude cold regions dramatically influence the formation and dissipation of clouds, affected by factors such as temperature, humidity, and wind speed [33,34]. This complexity complicates the spectral characteristics of clouds and snow, increasing the likelihood of cloud-snow confusion [35]. Additionally, the intricate terrain, characterized by ridges and valleys, results in significant variations in radiative properties across different areas [36,37], affecting the applicability of cloud detection algorithms.
Conversely, the plain or normal-temperature land in Target B present a more stable environment. This area experiences relatively minor climatic fluctuations, resulting in a simplified process of cloud formation and evolution, enhancing the consistency of spectral features between clouds and snow [38]. Moreover, the flatter terrain in the plains results in minimal variation in surface reflectance, improving the comparability of features across different regions and thereby enhancing the effectiveness of classification algorithms.

4.3. Performance Comparison with Existing Algorithms

As clearly illustrated in Figure 1, the operational cloud mask embedded within the AVHRR SR V5 product exhibits substantial overestimation. Consequently, it was excluded from this comparative analysis, which focuses on two established cloud detection algorithms specifically designed for AVHRR.

4.3.1. Overview of the Algorithm Performance Comparison

A quantitative intercomparison of the three algorithms is presented in Figure 7. Evaluation based on the average accuracy across cloud detection schemes reveals a significant performance enhancement achieved by the newly proposed algorithm. Compared to the methods introduced by Hori et al. and Hao et al., the improved algorithm demonstrates an increase in OA of 17.42% and 7.93%, and a notable increase in UA of 21.02% and 13.21%, respectively. These results highlight the main advantage of the improved algorithm, which is its effectiveness in alleviating the common problem of cloud overestimation in historical AVHRR data. The substantial gain in UA indicates a marked reduction in commission errors, meaning a significant decrease in false detection where bright surfaces like snow are misclassified as cloud. This effectively alleviates the inherent limitation of the AVHRR SR V5 product shown in Figure 1. The observed moderate decrease in PA by 8.12% and 9.21% reflects a deliberate trade-off in the algorithm design. By adopting a more conservative strategy that prioritizes the correct identification of clear-sky pixels, even at the cost of a slight increase in cloud omission, the algorithm provides a more reliable foundation for generating high-quality surface reflectance products. The concurrent improvement in the FS by 8.61% and 3.05% confirms that the algorithm achieves a superior balance between commission and omission errors. This balance is crucial for ensuring the continuity and reliability of long-term land surface parameter retrieval from time-series data.

4.3.2. Algorithm Performance Comparison of Different Detection Schemes

Further analysis categorized by schemes indicates that the performance gain is particularly pronounced in the high or cold land (Target A). Specifically, compared to the Hori algorithm, the OA for classes A1, A2, and A3 increased by 31.60%, 16.97%, and 43.42%, respectively, with FS improvements ranging from 7.23% to 16.27%. Compared to the Hao algorithm, OA improvements ranged from 5.06% to 43.34%, with FS gains between 2.46% and 9.76%. Figure 8 presents the newly developed optimal threshold determination scheme for BT3-BT4, along with the threshold distributions from two existing analogous algorithms. This comparative visualization provides an intuitive demonstration of the classification performance across the three methodologies. This notable advancement stems from the algorithm’s efficacy in mitigating cloud-snow confusion, a frequent challenge in high-latitude and mountainous regions. Traditional threshold-based methods often struggle with high-reflectance surfaces like snow, leading to high commission errors. The improvement was achieved by refining thresholds for the brightness temperature difference between the 3.7 µm and 11 µm channels, a critical spectral feature for cloud-snow discrimination, through the integration of spectral and topographic information.
The plain or normal-temperature land (Target B), the algorithm also demonstrates consistent and widespread improvement, as shown in Figure 9. The OA increased within ranges of 1.09–35.83% and 2.01–9.86% against the Hori and Hao algorithms, respectively, while FS improvements ranged from 0.54 to 17.51% and 0.46 to 5.13%. The most significant improvement within Target B was observed for class B1, where OA increased by 35.83% and 9.86%, and FS increased by 17.51% and 5.13% compared to the Hori and Hao algorithms, further validating the algorithm’s superiority in snow-/ice-covered scenarios. Furthermore, the newly developed classifiers for classes A4 and B6, which address surface types not effectively handled by the Hori algorithm, show improved OA of 6.21% and 3.14%, respectively, compared to the Hao algorithm, indicating broader applicability and enhanced robustness.

4.3.3. Algorithm Performance Comparison in Typical Regions

Figure 10 presents a cloud mask schematic of the optimized algorithm on 18 November 1988, along with a visual comparison of cloud detection results across four representative regions, such as the Alps, the Qinghai Tibet Plateau (QTP), Northeast China (NCN), and the Rocky Mountains (Rockies). For each region, the figure shows a false-color composite of the original AVHRR imagery alongside cloud masks generated by the original AVHRR SR V5 product, the Hori algorithm, the Hao algorithm, and the newly proposed multistage decision tree framework. The comparative visualization clearly demonstrates that the improved algorithm significantly reduces cloud overestimation over bright snow-covered surfaces, which is a persistent issue in the original AVHRR SR V5 product and remains partially unresolved in the two existing algorithms. In snow-abundant high-altitude regions such as the Tibetan Plateau, the Alps, and the Rocky Mountains, as well as frequently snow-covered areas of Northeast China, the new algorithm remains capable of accurately distinguishing snow from clouds, thereby minimizing commission errors while maintaining robust cloud detection. This enhanced discrimination capability stems from the refined thresholding of the BT3-BT4 brightness temperature difference, combined with integrated spectral and topographic contextual analysis, which improve the distinction between cool, bright snow and cloud.

5. Discussion

5.1. Analysis of Algorithm Uncertainty

5.1.1. Uncertainty of Truth Reference

Landsat 5 imagery, used as the truth reference for distinguishing between clouds and snow, has inherent limitations that may introduce significant uncertainty, particularly in complex scenarios (such as thin clouds and mixed pixels), where the cloud mask may suffer from omissions or misclassifications [39]. The morphology and distribution of clouds change rapidly over time, and the temporal difference between Landsat 5 and AVHRR may lead to inconsistencies in cloud cover. Although the improved “SNOMAP” algorithm achieves an accuracy exceeding 90% in forested and non-forested areas, limitations remain in regions with low snow cover (such as heterogeneous surfaces with uneven snow), where these errors may further propagate into the reference snow. Overall, despite some uncertainties, the algorithm’s overall accuracy remains satisfactory. Additionally, the significant differences in spatial scale between the high-resolution data of Landsat 5 and the low-resolution data of AVHRR may introduce errors during the scale transformation process.

5.1.2. Spectral Characteristic Overlap

Clouds and snow exhibit similar spectral characteristics in visible and near-infrared bands, particularly in the presence of thin clouds or shallow snow, where spectral overlap becomes more pronounced [40]. This similarity may lead to confusion in the algorithm’s ability to distinguish between clouds and snow. Furthermore, variations in atmospheric conditions (such as aerosols and water vapor) may exacerbate spectral similarity, especially in AVHRR data due to its lower spectral resolution [41].

5.1.3. Limitations of the Algorithm

The multi-threshold method relies on preset thresholds, which may not adapt well to complex surface conditions and variable atmospheric environments, particularly in areas with overlapping spectral characteristics or high surface heterogeneity [42]. Nevertheless, the algorithm demonstrates high accuracy in many scenarios.

5.2. Analysis of Application Prospects

5.2.1. Application and Characteristics of the Algorithm

This study initially identified overestimation of cloud detection due to cloud-snow spectral confusion in the AVHRR SR V5 product during the development of long-term snow cover products specifically for the Northern Hemisphere. To address this issue, an improved cloud detection algorithm was optimized for accuracy over snow-covered terrain using a multistage decision tree framework.
The algorithm employs a multistage decision tree architecture, significantly enhancing the differentiation between clouds and snow through a dynamic threshold optimization strategy. Quantitative assessments indicate that the overall accuracy of the new algorithm improves by 17.42% compared to the Hori algorithm and by 7.93% compared to the Hao algorithm. This improvement is particularly pronounced in areas with extensive snow cover in the Northern Hemisphere, effectively addressing the problem of snow surfaces being misclassified as clouds in the original product. By reducing the misclassification rate of snow surfaces, the algorithm effectively enhances the usability of cloud detection. Moreover, precise initial cloud detection reduces the computational burden of subsequent ground object recovery under the clouds.
Compared to machine Learning or deep learning, the advantages of this algorithm lie in its foundation as a traditional method based on physical rules and threshold optimization, offering strong interpretability and high computational efficiency. The algorithm can directly generate binary cloud masks, supports cloud-based processing on platforms such as Google Earth Engine without downloading raw spectral data, and can also be deployed locally, providing significant advantages in terms of time and storage costs. However, compared to deep learning, this method is somewhat less capable of capturing the complex nonlinear characteristics of cloud and snow pixels. Additionally, it is more sensitive to fluctuations in spectral features and faces challenges in threshold adjustment and generalization when applied across different sensors.

5.2.2. Extended Applications of Products

The improved cloud mask product demonstrates significant value in snow monitoring across the Northern Hemisphere. Snow cover detection primarily relies on optical remote sensing satellites [43], which are greatly affected by clouds. To obtain comprehensive surface snow information, it is necessary to conduct land cover recovery on remote sensing imagery, essentially performing cloud removal [44].
As illustrated in Figure 10a, the cloud mask product generated by the algorithm accurately represents the spatial distribution characteristics of cloud layers, providing a reliable data foundation for snow cover extraction. Moreover, the enhanced cloud mask has broad application potential in long-term time series of optical remote sensing parameter retrieval. For instance, in studies related to the Normalized Difference Vegetation Index (NDVI) [45], Leaf Area Index (LAI) [46], soil moisture [47], crop growth assessment [48], and land surface temperature product generation [49] in the Northern Hemisphere, the accurate cloud mask effectively differentiates cloud and non-cloud areas, significantly reducing the interference of clouds on remote sensing data and enhancing the accuracy of parameter retrieval. This not only optimizes snow monitoring and the preparation of related products but also provides more reliable data support for vegetation dynamics monitoring, agricultural resource management, and ecological environment assessments.

6. Conclusions

An enhanced multi-threshold cloud detection algorithm was developed for the AVHRR surface reflectance data to address the significant overestimation issue present in existing AVHRR cloud mask products, particularly due to cloud-snow spectral confusion. The algorithm was constructed and evaluated using Landsat 5 imagery as truth reference maps, where its inherent cloud mask and snow cover were identified by an improved “SNOMAP” algorithm, which served as cloud and snow samples, respectively. The model’s accuracy, uncertainties, and future development prospects were assessed.
The improved algorithm enhances cloud detection by adjusting the brightness temperature difference threshold between bands BT3 and BT4. Experimental results show an OA of 82.08%, with a PA of 86.35%, UA of 79.41%, and FS of 82.55%. Significant differences in detection accuracy were observed between high or cold land (Target A) and plains or normal-temperature land (Target B). Target A experienced greater cloud-snow confusion, particularly in subclass A3, which had a low OA of 67.43%. In contrast, Target B achieved a much higher OA of 90.93% in subclass B6 due to its environmentally stable and flat terrain.
Compared to the existing two algorithms, the improved algorithm effectively mitigates cloud overestimation in the AVHRR SR V5 product, showing improvements in OA by 17.42% and 7.93% and in UA by 21.02% and 13.21%. It achieves a better balance between commission and omission errors, indicated by FS enhancements of 8.61% and 3.05%. Optimizing the brightness temperature difference threshold between the 3.7 µm and 11 µm channels significantly reduces cloud-snow confusion in Target A, with a 43.42% OA improvement for Target A3. Target B also shows consistent performance, particularly with a 35.83% OA improvement for Target B1. The newly developed schemes for Targets A4 and B6 further extend the algorithm’s applicability, providing reliable technical support for generating high-quality, long-term cloud-free data products from AVHRR.
In summary, the enhanced cloud detection algorithm effectively mitigated the misclassification caused by cloud-snow spectral confusion in AVHRR data, significantly improving cloud detection accuracy, particularly over snow-covered regions in the Northern Hemisphere. Through a multi-level decision tree architecture and dynamic threshold optimization, the algorithm maintained high accuracy while enabling efficient computation, offering a robust solution for large-scale remote sensing data processing. The improved cloud mask product not only enhances snow cover monitoring quality but also provides a superior data foundation for applications such as vegetation analysis and agricultural management, holding significant importance for global environmental monitoring.

Author Contributions

Conceptualization, Q.Z. and X.H.; methodology, Q.Z., X.H. and D.S.; software, Q.Z., W.J. and Z.Z.; validation, Q.Z., W.J. and Z.Z.; resources, X.H. and D.S.; data curation, Q.Z. and X.H.; writing—original draft preparation, Q.Z.; writing—review and editing, X.H., D.S., W.J., G.H., Z.Z. and J.Z.; visualization, Q.Z. and W.J.; supervision, X.H.; funding acquisition, X.H., D.S., G.H. and J.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Science and Technology Department Program of Qinghai (No. 2024-ZJ-740), the National Natural Science Foundation of China (No. 42201153) and a grant from State Key Laboratory of Geographic Information Science and Technology.

Data Availability Statement

The sample data used in this study are available upon request.

Acknowledgments

The authors appreciate all the data provided by each open database. The authors would like to thank the editor and anonymous reviewers for their valuable comments and suggestions on this article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Yang, J.; Gong, P.; Fu, R.; Zhang, M.; Chen, J.; Liang, S.; Xu, B.; Shi, J.; Dickinson, R. The role of satellite remote sensing in climate change studies. Nat. Clim. Change 2013, 3, 875–883. [Google Scholar] [CrossRef]
  2. Chen, J.; Chen, S.; Fu, R.; Li, D.; Jiang, H.; Wang, C.; Peng, Y.; Jia, K.; Hicks, B.J. Remote sensing big data for water environment monitoring: Current status, challenges, and future prospects. Earth’s Future 2022, 10, e2021EF002289. [Google Scholar] [CrossRef]
  3. Pincus, R.; Hubanks, P.A.; Platnick, S.; Meyer, K.; Holz, R.E.; Botambekov, D.; Wall, C.J. Updated observations of clouds by MODIS for global model assessment. Earth Syst. Sci. Data 2023, 15, 2483–2497. [Google Scholar] [CrossRef]
  4. Prudente, V.H.R.; Martins, V.S.; Vieira, D.C.; de França e Silva, N.R.; Adami, M.; Sanches, I.D.A. Limitations of cloud cover for optical remote sensing of agricultural areas across South America. Remote Sens. Appl. Soc. Environ. 2020, 20, 100414. [Google Scholar] [CrossRef]
  5. Nolin, A.W. Recent advances in remote sensing of seasonal snow. J. Glaciol. 2010, 56, 1141–1150. [Google Scholar] [CrossRef]
  6. Huang, X.; Deng, J.; Ma, X.; Wang, Y.; Feng, Q.; Hao, X.; Liang, T. Spatiotemporal dynamics of snow cover based on multi-source remote sensing data in China. Cryosphere 2016, 10, 2453–2463. [Google Scholar] [CrossRef]
  7. Yamanouchi, T.; Suzuki, K.; Kawaguchi, S. Detection of clouds in Antarctica from infrared multispectral data of AVHRR. J. Meteorol. Soc. Japan. Ser. II 1987, 65, 949–962. [Google Scholar] [CrossRef]
  8. Hori, M.; Aoki, T.; Stamnes, K.; Chen, B.; Li, W. Preliminary validation of the GLI cryosphere algorithms with MODIS daytime data. Polar Meteorol. Glaciol. 2001, 15, 1–20. [Google Scholar]
  9. Karlsson, K.-G.; Stengel, M.; Meirink, J.F.; Riihelä, A.; Trentmann, J.; Akkermans, T.; Stein, D.; Devasthale, A.; Eliasson, S.; Johansson, E. CLARA-A3: The third edition of the AVHRR-based CM SAF climate data record on clouds, radiation and surface albedo covering the period 1979 to 2023. Earth Syst. Sci. Data 2023, 15, 4901–4926. [Google Scholar] [CrossRef]
  10. Stengel, M.; Stapelberg, S.; Sus, O.; Schlundt, C.; Poulsen, C.; Thomas, G.; Christensen, M.; Carbajal Henken, C.; Preusker, R.; Fischer, J. Cloud property datasets retrieved from AVHRR, MODIS, AATSR and MERIS in the framework of the Cloud_cci project. Earth Syst. Sci. Data 2017, 9, 881–904. [Google Scholar] [CrossRef]
  11. Kaps, A.; Lauer, A.; Kazeroni, R.; Stengel, M.; Eyring, V. Characterizing clouds with the CCClim dataset, a machine learning cloud class climatology. Earth Syst. Sci. Data 2024, 16, 3001–3016. [Google Scholar] [CrossRef]
  12. Barry, R.G.; Gan, T.Y. The Global Cryosphere: Past, Present, and Future; Cambridge University Press: Cambridge, UK, 2022. [Google Scholar]
  13. Villaescusa Nadal, J.L. Development of a Global Long Term Surface Albedo Data Record from Noaa Avhrr for the Estimation of 38 Year Trends (1982–2020). Diploma Thesis, University of Maryland, College Park, MD, USA, 2020. [Google Scholar]
  14. Hu, Z.; Kuenzer, C.; Dietz, A.J.; Dech, S. The potential of Earth observation for the analysis of cold region land surface dynamics in europe—A review. Remote Sens. 2017, 9, 1067. [Google Scholar] [CrossRef]
  15. Balsamo, G.; Agusti-Panareda, A.; Albergel, C.; Arduini, G.; Beljaars, A.; Bidlot, J.; Blyth, E.; Bousserez, N.; Boussetta, S.; Brown, A.; et al. Satellite and In Situ Observations for Advancing Global Earth Surface Modelling: A Review. Remote Sens. 2018, 10, 2038, Correction in Remote Sens. 2019, 11, 941.. [Google Scholar]
  16. Hori, M.; Sugiura, K.; Kobayashi, K.; Aoki, T.; Tanikawa, T.; Kuchiki, K.; Niwano, M.; Enomoto, H. A 38-year (1978–2015) Northern Hemisphere daily snow cover extent product derived using consistent objective criteria from satellite-borne optical sensors. Remote Sens. Environ. 2017, 191, 402–418. [Google Scholar] [CrossRef]
  17. Hao, X.; Huang, G.; Che, T.; Ji, W.; Sun, X.; Zhao, Q.; Zhao, H.; Wang, J.; Li, H.; Yang, Q. The NIEER AVHRR snow cover extent product over China—A long-term daily snow record for regional climate research. Earth System Science Data 2021, 13, 4711–4726. [Google Scholar] [CrossRef]
  18. Vermote, E. NOAA CDR Program. (2019): NOAA Climate Data Record (CDR) of AVHRR Surface Reflectance, Version 5; NOAA National Centers for Environmental Information: Asheville, NC, USA, 2019. [Google Scholar] [CrossRef]
  19. Wang, X.; Wang, J.; Che, T.; Huang, X.; Hao, X.; Li, H. Snow cover mapping for complex mountainous forested environments based on a multi-index technique. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 1433–1441. [Google Scholar] [CrossRef]
  20. Vermote, E.C.M. AVHRR Surface Reflectance and Normalized Difference Vegetation Index—Climate Algorithm Theoretical Basis Document; CDRP-ATBD-0459 Rev. 2; NOAA Climate Data Record Program: Asheville, NC, USA, 2018. [Google Scholar]
  21. Gorelick, N. Google Earth Engine. In EGU General Assembly Conference Abstracts; American Geophysical Union: Vienna, Austria, 2013; Volume 15, p. 11997. [Google Scholar]
  22. Crawford, C.J.; Roy, D.P.; Arab, S.; Barnes, C.; Vermote, E.; Hulley, G.; Gerace, A.; Choate, M.; Engebretson, C.; Micijevic, E. The 50-year Landsat collection 2 archive. Sci. Remote Sens. 2023, 8, 100103. [Google Scholar] [CrossRef]
  23. Foga, S.; Scaramuzza, P.L.; Guo, S.; Zhu, Z.; Dilley, R.D., Jr.; Beckmann, T.; Schmidt, G.L.; Dwyer, J.L.; Hughes, M.J.; Laue, B. Cloud detection algorithm comparison and validation for operational Landsat data products. Remote Sens. Environ. 2017, 194, 379–390. [Google Scholar] [CrossRef]
  24. Chen, S.; Wang, X.; Guo, H.; Xie, P.; Wang, J.; Hao, X. A conditional probability interpolation method based on a space-time cube for modis snow cover products gap filling. Remote Sens. 2020, 12, 3577. [Google Scholar] [CrossRef]
  25. Pan, F.; Jiang, L.; Wang, G.; Pan, J.; Huang, J.; Zhang, C.; Cui, H.; Yang, J.; Zheng, Z.; Wu, S. MODIS daily cloud-gap-filled fractional snow cover dataset of the Asian Water Tower region (2000–2022). Earth Syst. Sci. Data 2024, 16, 2501–2523. [Google Scholar] [CrossRef]
  26. DAAC, L. Global 30 Arc-Second Elevation Data Set GTOPO30; Land Process Distributed Active Archive Center: Sioux Falls, SD, USA, 2004. [Google Scholar]
  27. Carroll, M.; DiMiceli, C.; Townshend, J.; Sohlberg, R.; Hubbard, A.; Wooten, M. MOD44W: Global MODIS water maps user guide. Int. J. Digit. Earth 2017, 10, 207–218. [Google Scholar] [CrossRef]
  28. Ackerman, S.A.; Strabala, K.I.; Menzel, W.P.; Frey, R.A.; Moeller, C.C.; Gumley, L.E. Discriminating clear sky from clouds with MODIS. J. Geophys. Res. Atmos. 1998, 103, 32141–32157. [Google Scholar] [CrossRef]
  29. Stamnes, K.; Li, W.; Eide, H.; Aoki, T.; Hori, M.; Storvold, R. ADEOS-II/GLI snow/ice products—Part I: Scientific basis. Remote Sens. Environ. 2007, 111, 258–273. [Google Scholar] [CrossRef]
  30. Müller, R.; Pfeifroth, U. Remote sensing of solar surface radiation–a reflection of concepts, applications and input data based on experience with the effective cloud albedo. Atmos. Meas. Tech. 2022, 15, 1537–1561. [Google Scholar] [CrossRef]
  31. Pavolonis, M.J. Advances in extracting cloud composition information from spaceborne infrared radiances—A robust alternative to brightness temperatures. Part I: Theory. J. Appl. Meteorol. Climatol. 2010, 49, 1992–2012. [Google Scholar] [CrossRef]
  32. Pettorelli, N.; Vik, J.O.; Mysterud, A.; Gaillard, J.-M.; Tucker, C.J.; Stenseth, N.C. Using the satellite-derived NDVI to assess ecological responses to environmental change. Trends Ecol. Evol. 2005, 20, 503–510. [Google Scholar] [CrossRef]
  33. Beniston, M.; Diaz, H.; Bradley, R. Climatic change at high elevation sites: An overview. Clim. Change 1997, 36, 233–251. [Google Scholar] [CrossRef]
  34. Whiteman, C.D. Mountain Meteorology: Fundamentals and Applications; Oxford University Press: Oxford, UK, 2000. [Google Scholar]
  35. Houze, R.A., Jr. Cloud Dynamics; Academic Press: Cambridge, MA, USA, 2014; Volume 104. [Google Scholar]
  36. Wen, J.; Liu, Q.; Xiao, Q.; Liu, Q.; You, D.; Hao, D.; Wu, S.; Lin, X. Characterizing land surface anisotropic reflectance over rugged terrain: A review of concepts and recent developments. Remote Sens. 2018, 10, 370. [Google Scholar] [CrossRef]
  37. Shi, H.; Xiao, Z. Exploring topographic effects on surface parameters over rugged terrains at various spatial scales. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–16. [Google Scholar] [CrossRef]
  38. Fu, Y.; Ma, Y.; Zhong, L.; Yang, Y.; Guo, X.; Wang, C.; Xu, X.; Yang, K.; Xu, X.; Liu, L. Land-surface processes and summer-cloud-precipitation characteristics in the Tibetan Plateau and their effects on downstream weather: A review and perspective. Natl. Sci. Rev. 2020, 7, 500–515. [Google Scholar] [CrossRef]
  39. Zhu, Z.; Woodcock, C.E. Automated cloud, cloud shadow, and snow detection in multitemporal Landsat data: An algorithm designed specifically for monitoring land cover change. Remote Sens. Environ. 2014, 152, 217–234. [Google Scholar] [CrossRef]
  40. Stillinger, T.; Roberts, D.A.; Collar, N.M.; Dozier, J. Cloud masking for Landsat 8 and MODIS Terra over snow-covered terrain: Error analysis and spectral similarity between snow and cloud. Water Resour. Res. 2019, 55, 6169–6184. [Google Scholar] [CrossRef]
  41. King, M.D.; Kaufman, Y.J.; Menzel, W.P.; Tanre, D. Remote sensing of cloud, aerosol, and water vapor properties from the moderate resolution imaging spectrometer (MODIS). IEEE Trans. Geosci. Remote Sens. 1992, 30, 2–27. [Google Scholar] [CrossRef]
  42. Li, Y.; Wu, Y.; Li, J.; Sun, A.; Zhang, N.; Liang, Y. A Machine Learning Algorithm Using Texture Features for Nighttime Cloud Detection from FY-3D MERSI L1 Imagery. Remote Sens. 2025, 17, 1083. [Google Scholar] [CrossRef]
  43. Zhao, Q.; Hao, X.; Che, T.; Shao, D.; Ji, W.; Luo, S.; Huang, G.; Feng, T.; Dong, L.; Sun, X. Estimating AVHRR snow cover fraction by coupling physical constraints into a deep learning framework. ISPRS J. Photogramm. Remote Sens. 2024, 218, 120–135. [Google Scholar] [CrossRef]
  44. Hao, X.; Huang, G.; Zheng, Z.; Sun, X.; Ji, W.; Zhao, H.; Wang, J.; Li, H.; Wang, X. Development and validation of a new MODIS snow-cover-extent product over China. Hydrol. Earth Syst. Sci. 2022, 26, 1937–1952. [Google Scholar] [CrossRef]
  45. Chu, D.; Shen, H.; Guan, X.; Chen, J.M.; Li, X.; Li, J.; Zhang, L. Long time-series NDVI reconstruction in cloud-prone regions via spatio-temporal tensor completion. Remote Sens. Environ. 2021, 264, 112632. [Google Scholar] [CrossRef]
  46. Yan, K.; Park, T.; Yan, G.; Liu, Z.; Yang, B.; Chen, C.; Nemani, R.R.; Knyazikhin, Y.; Myneni, R.B. Evaluation of MODIS LAI/FPAR product collection 6. Part 2: Validation and intercomparison. Remote Sens. 2016, 8, 460. [Google Scholar] [CrossRef]
  47. Zhang, D.; Zhou, G. Estimation of soil moisture from optical and thermal remote sensing: A review. Sensors 2016, 16, 1308. [Google Scholar] [CrossRef]
  48. Doraiswamy, P.C.; Moulin, S.; Cook, P.W.; Stern, A. Crop yield assessment from remote sensing. Photogramm. Eng. Remote Sens. 2003, 69, 665–674. [Google Scholar] [CrossRef]
  49. Wu, P.; Yin, Z.; Zeng, C.; Duan, S.-B.; Göttsche, F.-M.; Ma, X.; Li, X.; Yang, H.; Shen, H. Spatially continuous and high-resolution land surface temperature product generation: A review of reconstruction and spatiotemporal fusion techniques. IEEE Geosci. Remote Sens. Mag. 2021, 9, 112–137. [Google Scholar] [CrossRef]
Figure 1. Schematic of (a) false-color imagery (SREFL_CH3, CH2, CH1) and (b) cloud mask in the NOAA AVHRR CDR SR V5 product on 18 November 1988, showing evident cloud overestimation.
Figure 1. Schematic of (a) false-color imagery (SREFL_CH3, CH2, CH1) and (b) cloud mask in the NOAA AVHRR CDR SR V5 product on 18 November 1988, showing evident cloud overestimation.
Remotesensing 17 03992 g001
Figure 2. The Landsat 5 reference SCE generation process for a single sample.
Figure 2. The Landsat 5 reference SCE generation process for a single sample.
Remotesensing 17 03992 g002
Figure 3. The Landsat 5 reference cloud generation process for a single sample.
Figure 3. The Landsat 5 reference cloud generation process for a single sample.
Remotesensing 17 03992 g003
Figure 4. The framework for cloud detection algorithm improvement [16,17].
Figure 4. The framework for cloud detection algorithm improvement [16,17].
Remotesensing 17 03992 g004
Figure 5. Framework of the existing cloud detection algorithm for the Northern Hemisphere.
Figure 5. Framework of the existing cloud detection algorithm for the Northern Hemisphere.
Remotesensing 17 03992 g005
Figure 6. Accuracy evaluation results of the ten cloud detection schemes.
Figure 6. Accuracy evaluation results of the ten cloud detection schemes.
Remotesensing 17 03992 g006
Figure 7. Accuracy evaluation result comparison between the new improved algorithm and existing two algorithms.
Figure 7. Accuracy evaluation result comparison between the new improved algorithm and existing two algorithms.
Remotesensing 17 03992 g007
Figure 8. Frequency distribution of BT3-BT4 and optimal thresholds for cloud detection comparison with two existing algorithms in four schemes in high-altitude or cold land [16,17].
Figure 8. Frequency distribution of BT3-BT4 and optimal thresholds for cloud detection comparison with two existing algorithms in four schemes in high-altitude or cold land [16,17].
Remotesensing 17 03992 g008
Figure 9. Frequency distribution of BT3-BT4 and optimal thresholds for cloud detection comparison with two existing algorithms in the plains or normal-temperature land [16,17].
Figure 9. Frequency distribution of BT3-BT4 and optimal thresholds for cloud detection comparison with two existing algorithms in the plains or normal-temperature land [16,17].
Remotesensing 17 03992 g009
Figure 10. (a) Demonstration of cloud masks identified by the improved algorithm using NOAA AVHRR CDR V5 data (18 November 1988); (b) Intercomparison of original false-color imagery and cloud masks (original product, new algorithm, Hori algorithm, Hao algorithm) in the Alps, the Qinghai Tibetan Plateau (QTP), Northeast China (NCN), and the Rocky Mountains (Rockies).
Figure 10. (a) Demonstration of cloud masks identified by the improved algorithm using NOAA AVHRR CDR V5 data (18 November 1988); (b) Intercomparison of original false-color imagery and cloud masks (original product, new algorithm, Hori algorithm, Hao algorithm) in the Alps, the Qinghai Tibetan Plateau (QTP), Northeast China (NCN), and the Rocky Mountains (Rockies).
Remotesensing 17 03992 g010
Table 1. AVHRR SR V5 product band information.
Table 1. AVHRR SR V5 product band information.
Band NameUnitsScaleWavelengthDescription
SR1 0.0001640 nmBidirectional surface reflectance
SR2 0.0001860 nmBidirectional surface reflectance
SR3 0.00013.75 umBidirectional surface reflectance
BT3K0.13.75 umBrightness temperature
BT4K0.111.0 umBrightness temperature
BT5K0.112.0 umBrightness temperature
RAAdeg0.01 Relative sensor azimuth angle
SZAdeg0.01 Solar zenith angle
VZAdeg0.01 View zenith angle, scale 0.01
QA Quality control bit flags
Table 2. AVHRR SR V5 product QA bitmask information.
Table 2. AVHRR SR V5 product QA bitmask information.
Bitmask ValueDescription01
Bit 0UnusedNoYes
Bit 1Pixel is cloudyNoYes
Bit 2Pixel contains cloud shadowNoYes
Bit 3Pixel is over waterNoYes
Bit 4Pixel is over sunglintNoYes
Bit 5Pixel is over dense dark vegetationNoYes
Bit 6Pixel is at night (high solar zenith)NoYes
Bit 7Channels 1–5 are validNoYes
Bit 8Channel 1 value is invalidNoYes
Bit 9Channel 2 value is invalidNoYes
Bit 10Channel 3 value is invalidNoYes
Bit 11Channel 4 value is invalidNoYes
Bit 12Channel 5 value is invalidNoYes
Bit 13RHO3 value is invalidNoYes
Bit 14BRDF correction is invalidNoYes
Bit 15Polar flag, latitude over 60 degrees (land) or 50 degrees (ocean)NoYes
Table 3. Landsat 5 SR product band information.
Table 3. Landsat 5 SR product band information.
NameUnitsScaleWavelength (μm)Spatial ResolutionDescription
SR_B1 2.75 × 10−50.45~0.5230 mBand 1 (blue) surface reflectance
SR_B2 2.75 × 10−50.52~0.6030 mBand 2 (green) surface reflectance
SR_B3 2.75 × 10−50.63~0.6930 mBand 3 (red) surface reflectance
SR_B4 2.75 × 10−50.77~0.9030 mBand 4 (near-infrared) surface reflectance
SR_B5 2.75 × 10−51.55~1.7530 mBand 5 (shortwave infrared 1) surface reflectance
ST_B6K3.418 × 10−310.40~12.50120 mBand 6 surface temperature.
SR_B7 2.75 × 10−52.08~2.3530 mBand 7 (shortwave infrared 2) surface reflectance
QA_PIXEL Pixel quality attributes generated from the CFMASK algorithm
Table 4. Landsat 5 SR product QA bitmask information.
Table 4. Landsat 5 SR product QA bitmask information.
Bitmask ValueDescription012
Bit 0Fill
Bit 1Dilated Cloud
Bit 2Unused
Bit 3Cloud
Bit 4Cloud Shadow
Bit 5Snow
Bit 6ClearCloud or Dilated Cloud bits are setCloud and Dilated Cloud bits are not set
Bit 7Water
Bit 8–9Cloud ConfidenceNoneLowMedium
Bit 10–11Cloud Shadow ConfidenceNoneLowMedium
Bit 12–13Snow/Ice ConfidenceNoneLowMedium
Bit 14–15Cirrus ConfidenceNoneLowMedium
Table 5. The multi-level decision tree cloud detection algorithm input parameters.
Table 5. The multi-level decision tree cloud detection algorithm input parameters.
ClassificationParameter
Surface reflectance data (SR)SR1SR2SR3
Bright temperature data (BT)BT4
Index and band combination data (I&B)NDVISR1-SR2BT3-BT4BT4-BT5
Terrain data (TD)DEM
Table 6. Confusion matrix.
Table 6. Confusion matrix.
Confusion MatrixPrediction/Product
Reference PositiveNegative
PositiveTPFN
NegativeFPTN
Table 7. Improved multi-level decision tree cloud detection algorithm and thresholds.
Table 7. Improved multi-level decision tree cloud detection algorithm and thresholds.
TargetTarget Serial NumberSwitchElevation
(m)
SR1SR2SR3SR1-SR2NDVIBT4BT3-BT4BT4-BT5
A: high or cold land
(DEM > 300 and BT4 < 260 K)
A1On<3000 ≥240>20
A2On≥3000 ≥240>23.6
A3On <240>32.8
A4On >0.1>0.02 >31.4
B: plains or normal-temperature land
(other land)
(DEM < 300 and BT4 ≥ 260 K)
B1On <260>21.4
B2On >−0.02 <310>16
B3On >0.3 >−0.02 <293>16
B4On >0.4 >−0.03 <293>16.8>−1
B5On >0.4 <278>16.4>−1
B6On >0.3 >0.02 >16.4
B7Off >0.5>288
B8Off >310
B9Off>1000<0.4 <−0.04 >275
B10Off <−0.04 >300
1 BT3-BT4 are the primary improved band thresholds.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhao, Q.; Hao, X.; Shao, D.; Ji, W.; Huang, G.; Zhao, Z.; Zhang, J. Optimizing Cloud Mask Accuracy over Snow-Covered Terrain with a Multistage Decision Tree Framework. Remote Sens. 2025, 17, 3992. https://doi.org/10.3390/rs17243992

AMA Style

Zhao Q, Hao X, Shao D, Ji W, Huang G, Zhao Z, Zhang J. Optimizing Cloud Mask Accuracy over Snow-Covered Terrain with a Multistage Decision Tree Framework. Remote Sensing. 2025; 17(24):3992. https://doi.org/10.3390/rs17243992

Chicago/Turabian Style

Zhao, Qin, Xiaohua Hao, Donghang Shao, Wenzheng Ji, Guanghui Huang, Zisheng Zhao, and Juan Zhang. 2025. "Optimizing Cloud Mask Accuracy over Snow-Covered Terrain with a Multistage Decision Tree Framework" Remote Sensing 17, no. 24: 3992. https://doi.org/10.3390/rs17243992

APA Style

Zhao, Q., Hao, X., Shao, D., Ji, W., Huang, G., Zhao, Z., & Zhang, J. (2025). Optimizing Cloud Mask Accuracy over Snow-Covered Terrain with a Multistage Decision Tree Framework. Remote Sensing, 17(24), 3992. https://doi.org/10.3390/rs17243992

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop