Next Article in Journal
A Multi-Pixel Split-Window Approach to Sea Surface Temperature Retrieval from Thermal Imagers with Relatively High Radiometric Noise: Preliminary Studies
Previous Article in Journal
QMRNet: Quality Metric Regression for EO Image Quality Assessment and Super-Resolution
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Scale Conversion Model Based on Deep Learning of UAV Images

1
National Engineering Laboratory for Satellite Remote Sensing Applications, Aerospace Information Research Institute, Chinese Academy of Sciences, Beijing 100101, China
2
University of Chinese Academy of Sciences, Beijing 100049, China
3
College of Earth Sciences, Hebei Geo University, Shijiazhuang 050030, China
4
College of Geo-Exploration Science and Technology, Jilin University, Changchun 130026, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(9), 2449; https://doi.org/10.3390/rs15092449
Submission received: 10 April 2023 / Revised: 25 April 2023 / Accepted: 27 April 2023 / Published: 6 May 2023

Abstract

:
As a critical component of many remote sensing satellites and model validation, pixel-scale surface quantitative parameters are often affected by scale effects in the acquisition process, resulting in deviations in the accuracy of image scale parameters. Consequently, various successive scale conversion methods have been proposed to correct the errors caused by scale effects. In this study, we propose ResTransformer, a deep learning model for scale conversion of surface reflectance using UAV images, which fully extracts and fuses the features of UAV images in the sample area and sample points and establishes a high-dimensional nonlinear spatial correlation between sample points and sample area in the target sample area, so that the scale conversion of surface reflectance at the pixel-scale can be completed quickly and accurately. We collected and created a dataset of 500k samples to verify the accuracy and robustness of the model with other traditional scale conversion methods. The results show that the ResTransformer deep learning model works best, providing average MRE, average MRSE, and correlation coefficient R values of 0.6440%, 0.7460, and 0.99911, respectively, and the baseline improvements compared with the Simple Average method are 92.48%, 92.45%, and 16.59%, respectively. The ResTransformer model also shows the highest robustness and universality and can adapt to surface pixel-scale conversion scenarios with different sizes, heterogeneous sample areas, and arbitrary sampling methods. This method provides a promising, highly accurate, and robust method for converting pixel-scale surface reflectance scale.

Graphical Abstract

1. Introduction

Scale conversion is a critical link in many remote sensing physical modelling, remote sensing product applications and quantitative description of surface parameters at the pixel scale [1,2,3,4]. The primary problem in remote sensing scale conversion is effectively converting remote sensing data and information from one scale to another and simultaneously giving the evaluation index and uncertainty of the scale conversion results [2,5,6]. The pixel-scale surface reflectance is necessary for satellite verification and inversion of other quantitative remote sensing parameters [7]. In obtaining the surface reflectance at the pixel scale, the surface reflectance of the sampling points collected on the ground needs to be scaled up to the pixel scale via the point-to-surface scale conversion method [8,9,10].
The main methods of upscaling remote sensing parameters in the past decades include Simple Average, empirical regression, geostatistical, and Bayesian methods. Among them, the Simple Average method calculates the arithmetic mean of multiple sample points or the weighted mean of each sample point according to the components within the sample area as the actual image element scale value, which is a simple method [11,12,13]. However, it only applies to areas with low heterogeneity of surface parameters. Otherwise, the results will bear significant uncertainty [14]. The empirical regression method establishes the interrelationship between the measured values of each sample point, the related metadata and the actual values of the image element scale through the historical observation samples. A series of regression models are then used such as the least squares method to fit [15,16,17], which has improved accuracy compared with Simple Average, but is overly dependent on historical observation data and has poor portability [18]. The geostatistical method aims to combine the theory of regional variation with the variation function as a tool to reconstruct the measured values of internal sampling points to the metric scale [19,20]. Moreover, the geostatistical method has no interference from other metadata, and its accuracy is only related to the distribution of the number of sampling points; however, this also leads to an overdependence of its result accuracy on the sampling method [21]. The Bayesian method adopts the Bayesian maximum entropy method [22], incorporates various prior information, and has no requirement for the distribution of prior knowledge [23]. However, this method is more complex, and the probabilistic transformation of various prior knowledge is still in the research stage.
Due to subsurface heterogeneity during the field measurement of quantitative remotely sensed surface parameters, the quantitative inversion models of the corresponding parameters and the accuracy of the results produce variations of different sizes. While validating the lake surface albedo model, Du et al. found that spatial heterogeneity led to lower precision of the point-scale reflectance due to the lack of scale conversion, which significantly impacted the accuracy of the validation results of the model [24]. The current research on scale conversion for many remotely sensed surface parameters mainly focuses on surface temperature, LAI, BRDF, total primary productivity and landscape phenology parameters. Liang et al. used high-resolution remote sensing images to upscale ground point-scale reflectance measurements to low-resolution images with reasonable accuracy [25]. Yue et al. used high-resolution remote sensing images to upscale ground point-scale albedo measurements to the 500 m image scale [26], and Sanchez-Zapero et al. combined the spatial heterogeneity of the subsurface. They employed geostatistical methods that combined the subsurface’s spatial heterogeneity and scaled the observation tower’s albedo measurements to the image element scale by the geostatistical method and the Simple Average method [27].
Due to the influence of surface heterogeneity, sampling method and randomness of sampling in the image element range, the scale conversion of the traditional point surface conversion method yields low accuracy of the pixel-scale surface reflectance, and the uncertainty of the scale conversion results fluctuates wildly [28]. For example, the traditional Simple Average method provides high accuracy of the image element-scale surface reflectance for sites with low surface heterogeneity, while the uncertainty of the scale conversion results is more significant for areas with high heterogeneity. Regarding the empirical regression method, due to the convenience of ground truth surface reflectance measurements, and based on the principle of incorporating as many types of features and multiple sample areas as possible [29], the sampling locations are often irregular and random, making it challenging to obtain historical observation samples [30]. The surface reflectance will change even in the same area, so the empirical model obtained in the previous period does not apply to scale conversion of the current environment. The geostatistical and general interpolation method requires a high number and distribution of sample points. It often performs better in areas with a high density of sample points. However, because the accessibility of different feature types in the same area differs, the number of sample points may be small and the distribution of sample points may be uneven [31]. Hence, the accuracy of the surface reflectance results obtained by the geostatistical method varies significantly in different types of sample areas, and the method’s robustness is low. In summary, the traditional point-to-surface conversion method is unsuitable for pixel-scale surface reflectance acquisition.
Therefore, this paper proposes an adaptive and highly robust ResTransformer surface-reflectance scale-conversion deep learning model that combines UAV images. The model extracts, fuses, and adaptively learns the features within the UAV image of the target sample area and the UAV image of the sampling points within the sample area through the ResTransformer deep learning model. It then establishes a high-dimensional nonlinear spatial correlation between the sampling points within the target sample area and the sample area, and maps the measured surface reflectance of each sampling point within the sample area to the surface reflectance of the sample area through the ResTransformer model. The surface reflectance of each sampling point within the sample area is mapped to the surface reflectance of the sample area by the ResTransformer model. It is possible to scale the surface reflectance of arbitrary size, heterogeneous sample area, arbitrary sampling location and the number of scenes with high-precision pixel scale.
The main objective of this study is to develop a deep learning model with UAV images to achieve high-accuracy pixel-scale surface-reflectance scale conversion for arbitrary size, heterogeneous sample area, arbitrary sampling location and the number of scenes. This method provides a relatively easy-to-use method to scale the surface reflectance, and the average MRE, average MRSE, and correlation coefficient R with the actual values on the large-scale dataset being 0.6440%, 0.7460 and 0.99911, respectively, 92.45% and 16.59%, respectively, compared to the baseline method. It is also found that the scale conversion accuracy of the deep learning model remains high for arbitrary size, heterogeneous sample area, arbitrary sampling location and the number of scenarios. Moreover, it demonstrates higher robustness and generalizability than the traditional methods. Section 2 of this paper introduces the normalization method for the reflectance scale conversion problem and the data source and construction of the large-scale reflectance scale conversion dataset. Section 3 introduces the ResTransformer deep learning model and the evaluation method of the accuracy of the scale conversion results. Section 4 analyzes and evaluates the accuracy of various scale conversion models on the dataset, and we conclude the above work in Section 5.

2. Reduction Method and Dataset

2.1. Reduction Method

Reduction refers to converting a complex problem or expression into a more straightforward form that can be solved or processed more efficiently [32]. It is commonly used in computer science for solving problems in computational theory, such as proving that a problem is NP-complete and for transforming a complex problem into a known simple problem to be solved [33,34]. The prerequisites for reductio ad absurdum are that problem A can be transformed into problem B via rules and algorithms, and that both problems have the same type of answer, and the reduction is reversible and valid [35,36].
Since the process of imaging the sample area observation by UAV is similar to the process of detecting the reflectance of the sample point by the spectrometer above the sample point [37,38,39], the scale conversion problem of surface reflectance is considered to be reduced to the scale conversion problem of the grayscale value of the sampling point of the UAV image, as shown in Figure 1. The surface reflectance at the image element scale for any sample area N is the average of the reflectance of each continuous element within it [40]. However, measuring the reflectance of each continuous element in the field measurement is impossible, and the efficiency could be better [41]. Therefore, only the reflectance  R e f ( n k )  of k sampling points inside the sample area N is usually measured, and the image-scale surface reflectance of the sample area is obtained by establishing a functional relationship  F  between the reflectance of the sampling points and the reflectance  R e f ( N )  of the sample area N, as shown in Equation (1).
R e f ( N ) = F ( R e f ( n 1 ) , R e f ( n 2 ) , , R e f ( n k ) )
The panchromatic images obtained from UAV capture can simultaneously measure the grayscale values of each consecutive element within the sample area N. Therefore, the complex nonlinear correlation between all the sampling points in the sample area N and the sample area N is established by using the UAV images in the sample area, all the UAV images in the sampling points, and the average grey value  D N ( N )  of the UAV images in the sample area and the average grey value  D N ( n k )  of the UAV images in the sampling points are calculated to optimize and verify F, as shown in Equation (2).
D N ( N ) = F ( D N ( n 1 ) , D N ( n 2 ) , , D N ( n k ) )
Moreover, the result of the convolution of the surface reflectance of the sample area and the spectral response function of the UAV imaging spectrometer is related to the grayscale value of the UAV image so that Equation (3) exists, allowing arbitrary surface reflectance to be converted to the grayscale value of the UAV image in polynomial time, where  Γ ( λ )  is the spectral response function at the corresponding wavelength.
D N ( N ) = 0 R e f N ( λ ) Γ ( λ ) d λ
In summary, the input of the pixel-scale surface reflectance scale conversion is converted into the input of the scale conversion of the UAV image of the sampling point using Equation (3). The features of the sample area and the UAV image of the sampling point are extracted by the deep learning model to obtain the feature representation of the corresponding space. The similarity weights of the features of the sample area and the sampling point are then obtained by feature fusion. After that, the average DN value of the UAV images of the sampling points is multiplied by the corresponding feature weights, and the corresponding corrected sampling results are obtained through the fully connected layer calculation. The nonlinear correlation model between the UAV images of the sampling area and the UAV images within the sampling points is then established by mining the complex nonlinear spatial relationship between the effective sampling points and the sample area, solving the problem of scale conversion of the grayscale values of the sampling points of the UAV images, and the correlation model is applied to the reflectance of the sampled points to complete the normalization of the reflectance scale conversion problem.

2.2. UAV Image Data

A DJI (DJI Technology Co., Ltd., Shenzhen, China) Mavic 2 UAV took the UAV images used in this paper with a diagonal field of view of 84° and a single shot of 3000 px × 4000 px in actual size. Our team travelled to Xiong’an, Hebei, Shihezi, Xinjiang, Yuanmou, Yunnan, and Sanya, Hainan, between 2022 and 2023 to capture feature types specific to the destinations. Figure 2 shows the image locations and feature types. Table 1 shows the number of UAV images of each feature type and the observation area corresponding to the UAV images.

2.3. Dataset Construction

A large number of a priori samples are needed to extract the features and correlation between the corresponding sample areas and the sampled UAV images by deep learning models and to solve the scale conversion problem of the grayscale values of the sampled UAV images, for which the construction of a large-scale dataset is one of the critical steps [42,43]. In this paper, we build a large-scale scale-conversion dataset (SCD) containing multiple feature types, sample area sizes and sampling methods based on the UAV images acquired from several field experiments, and it includes 500k training samples. Figure 3 shows the overall process of SCD dataset construction.
The UAV images obtained from the field experiment acquisition first need to judge whether the image acquisition height is greater than 2 m to avoid the sample area being too small, leading to many overlapping simulated sampling points, which affects the accuracy of the data set. Since the surface reflectance acquisition and the pixel-scale surface reflectance acquisition process require the ground sample area to be square, and the UAV image is a rectangle of 3000 px × 4000 px, so for any UAV image, it needs to be cropped to a square of 3000 px in length and width. The SCD database is a large-scale pixel-scale reflectance scale conversion dataset containing multiple feature types, multiple sample area sizes, and multiple sampling methods. In addition, the deep learning model requires a large number of a priori samples, in order to increase the training samples, perform sample enhancement and achieve the purpose of simulating different high-spatial-resolution satellite pixel-scale size sample area scale conversion, by simulating UAV at different altitudes. The sample enhancement is achieved by simulating the UAV at different heights and imaging the sample area to obtain a variety of samples with different sample sizes. Figure 4 illustrates the overall process of sample enhancement. The spatial resolution of individual pixels of the original UAV image is calculated by Equation (4). The edge length of the square sample area under the simulated height h is calculated by Equation (5), in which  2 htan ( 42 ° )  is the length of the diagonal of the corresponding enhanced sample because the cutting according to the long edge is prone to the problem of wide edge crossing, so the short edge cuts the enhanced sample as the square sample area. The short edge length is 0.6 times the diagonal length, and the corresponding sample area is calculated by Equation (6). The number of pixel points in the edge length of the enhanced sample simulated UAV image  w i d t h p x  is calculated by Equation (7).
resolution = 2 Htan ( 42 ° ) × 3 5 3000 = Htan ( 42 ° ) 2500
w i d t h = 2 htan ( 42 ° ) × 3 5
a r e a = w i d t h × w i d t h = 36 25 h 2 tan 2 ( 42 )
w i d t h p x = w i d t h r e s o l u t i o n = 6 5 htan ( 42 ° ) Htan ( 42 ° ) 2500 = 3000 h H
After obtaining the UAV images of the corresponding sample area through the above steps, it is necessary to simulate the field feature spectral acquisition process by selecting sampling points on the UAV images for simulated sampling. Sampling point selection includes standard and random sampling, and standard sampling adopts equal spacing sampling. The sampling methods include one-point sampling, two-point sampling, 2 × 2 four-point sampling, five-point sampling, 3 × 3 nine-point sampling and 4 × 4 sixteen-point sampling [44]. Random sampling uses randomly selected points, and the number of sampling points ranges from 1 to 16. Figure 5 shows the sampling locations of various sampling methods.
In order to make the samples in the dataset as close as possible to the actual field reflectance acquisition scene, we limit the acquisition height of each sampling point to 1 m~1.4 m by randomly selecting the sampling height  h point  as the sampling point. The side length  w point  of the corresponding sampling point UAV image is calculated by Equation (6). The corresponding sampling radius  r point  can be calculated by Equation (7), where  I F O V S E I  is the SEI spectrometer of IFOV. The IFOV in SCD was set to 25° because the field of view of the fiber optic probe of most of the mainstream spectrometers, such as SEI and ASD, is 25°.
w point = 5000 h point tan ( I F O V S E I ) h U A V tan ( 42 )
r point = h point tan ( I F O V S E I ) h U A V tan ( 42 )
After obtaining the sample area UAV image and the sample location by simulated cropping and simulated sampling, the UAV image of the sample location is cropped. The average DN values of the sample point and the sample area UAV grayscale images are calculated, respectively, to complete the construction of the single SCD sample. Figure 6 shows the sample construction process. Since then, a single sample in the dataset has been created, which contains the UAV images of the sample area and the sample point and the average DN values of the UAV grayscale images of the sample area and the sample point.
Since the ResTransformer model is not an end-to-end model, it requires the same input data shape. However, in the process of surface reflectance measurement, due to various natural factors such as topography, geomorphology and subsurface state, the size of each sample area, the number, location and sampling height of sampling points are different, resulting in different resolutions of UAV images of sampling points within the sample area and the number of average DN values of sampling points. Hence, it is necessary to unify the samples in the SCD dataset by coding each. The samples in the SCD dataset need to be coded into the same structure.
Each sample in the SCD dataset includes the UAV images of the sample area, the sample points, and the DN mean of the UAV grayscale images. At the same time, for the UAV images of the sample points, since the number of sample points is usually less than 16 in the ground-based spectral acquisition process, we reserve 16 filled areas for the sample points and place the sample point images into the squares along the image channel dimension, each of which has a size of 56 px × 56 px. Similarly, the average grey value of the sampled UAV images is filled into a matrix of length 16 by cyclic mode-taking, and Figure 7 illustrates the sample normalization process.

3. Methods and Evaluation Indicators

3.1. Traditional Scale Conversion Methods

The traditional scale conversion methods include Simple Average, Ordinary Kriging, and Cubic Spline Interpolation. Among them, the Simple Average method uses Equation (8) to calculate the mean value of DN values of each sampling point in the sample area as the scale conversion result of the sample area. Where n is the number of sample points in the SCD data set,  D N point _ i  is the average DN value of the UAV grayscale image at the ith sample point.
r e s u l t s i m p l e _ a v e r a g e = 1 n i = 1 n D N point _ i
The Ordinary Kriging method first creates a variance function and a covariance function to estimate the spatial autocorrelation value. In this paper, a Gaussian function is used as the variance function. Equation (9) shows the mathematical model of the Ordinary Kriging method, where  Z # ( x )  is the unbiased optimal estimate of the DN value of the UAV grayscale image at location (x,y).  Z i ( a = 1 , 2 , · · · n )  is the mean value of DN of the UAV grayscale image at the sampling point. According to the two conditions of unbiased and optimality, the estimation equation of the weight coefficient  λ i  is shown in Equation (10). The solution of the equation shown in Equation (10) is brought into Equation (9) to obtain the estimated value of DN at the coordinates (x,y). Finally, the scale conversion result of the sample is obtained by calculating the mean value of all DN values within the sample area [45,46,47].
Z # ( x ) = i = 1 N λ i Z i
β = 1 N λ β γ ( x α , x β ) + L = 0 K μ L f L ( x α ) + μ 0 = r ( x 0 , x ) α = 0 N λ α f L ( x α ) = f L ( x ) β = 1 N λ β γ ( x α , x β )
The cubic Spline Interpolation method is a smooth surface interpolation method for estimating the function values between given data points. Similar to the one-dimensional case, it divides the surface between a given data point into multiple cubic function slices and ensures that the function and derivative values between adjacent slices are equal, thus ensuring the continuity and smoothness of the surface. Equation (11) shows the general form of Cubic Spline Interpolation, where  D N ( x , y )  is the DN value of the whole sample plane,  w i , j  is the weight,  N i ( x ) N j ( y )  is the basis function, and m and n are the total numbers of sampled points in the length and width direction of the sample area, respectively. The basis function is usually three-times the spline function; they have good smoothness and approximation properties. Equation (12) defines the spline function, where  h i  is the distance between two sampling points [48,49,50].
D N ( x , y ) = i = 1 m j = 1 n w i , j N i ( x ) N j ( y )
N i ( x ) = 1 6 h i x x i 1 3 , x i 1 x < x i 3 h i 2 x x i 1 2 3 h i x x i 1 3 + h i 3 , x i x < x i + 1 3 h i 2 x i + 1 x 2 3 h i x i + 1 x 3 + h i 3 , x i + 1 x < x i + 2 x i + 2 x 3 , x i + 2 x

3.2. ResTransformer Deep Learning Model

The ResTransformer model proposed in this paper is a deep learning framework which aims to scale the conversion of surface reflectance by extracting the features of the UAV images of the learning sampling points and the UAV images of the sample area and establishing the correlation between the average DN values of the UAV images of the sampling points and the average DN values of the UAV images of the sample area. The ResTransformer model consists of Resnet and Swim Transformer V2. ResNet is a deep residual network that introduces the residual block concept to add input and output data to learn residuals [51]. This residual learning approach can solve the gradient disappearance or gradient explosion problem encountered by neural networks during training and is widely used in computer vision tasks. Transformer is a deep learning model based on a self-attentive mechanism, initially proposed by Vaswani et al. in 2017 [52], whose core idea is encoding and decoding sequences through a multi-headed attention mechanism. The Transformer model first achieved excellent results in natural language processing [53,54]. With the Vision proposed by Dosovitskiy et al. in 2020 to achieve an image serialization Transformer model, Transformer has also been gradually applied to computer vision [55]. A successive series of Transformer-based computer vision models have been proposed, such as DETR and Swin Transformer [56].
The ResTransformer first extracts the features of the sample area and the sample point UAV images, obtains the spatial representation of the corresponding features and then uses feature fusion to obtain the similarity weights  λ  of the sample area and the sample point features. The average DN value of the sample points multiplies the corresponding feature weight  λ  to obtain the corresponding corrected sampling results and establish the complex non-linear spatial relationship  f  between the effective sample points and the sample area, after which the corresponding scale conversion results are obtained by Equation (13).
r e s u l t ResTransformer = f ( λ 1 D N 1 , λ 2 D N 2 , · · · λ n D N n )
ResTransformer combines the advantages of ResNet and Transformer, using ResNet-18 as the backbone network for feature extraction of sample area and sample point UAV images. We use a double-headed Resnet Block to extract features from the sample area and sample point UAV images. The sample area image size is 224 × 224 × 3, so 12 Resnet Blocks are used to compress the feature size to 14 × 14 × 256 after four times of downsampling, and the sample point UAV image size is compressed to 14 × 14 × 64 by 4 Resnet Blocks after two times of downsampling, after which the sample area and sample point UAV image features are stitched according to the channel dimension.
The above output image features are passed through four Resnet Blocks and fused after one downsampling. The typical features are then extracted and computed using an attention mechanism by the Swin Transformer Block V2 module [57]. The input UAV image features in Swin Transformer Block V2 are split into multiple patches, each transformed independently. The different patches are connected by Cross-Stage Connection (CSC) and Local Window Swapping (LWSW). The information is exchanged and integrated between the patches through CSC and LWS. After fusing the output features of the encoder with the input of the linear layer, the complex non-linear spatial relationship between the sample area and the sampled points is obtained through a multi-layer full connection and a non-linear combination of activation functions.
The above output image features are passed through four Resnet Blocks and fused after one downsampling. After that, the typical features are extracted and computed by the Swin Transformer Block V2 module using the attention mechanism. The input UAV image features in Swin Transformer Block V2 are partitioned into multiple patches, and each patch is independently transformer computed. The different patches are connected by Cross-Stage Connection (CSC) and Local Window Swapping (LWSW). The information is exchanged and integrated between different patches through CSC and LWS. After fusing the output features of the encoder with the input of the linear layer, the complex nonlinear spatial relationship between the sample area and the sampling points is obtained by the nonlinear combination of the multilayer full connection and the activation function, and the scale transformation results are obtained by combining the above Equation (13). Figure 8 illustrates the overall architecture of ResTransformer.
To quantitatively describe the computational complexity of the ResTransformer model, we calculated the Floating-Point Operations (Flops) for each module of the ResTransformer model and the parameters of the corresponding modules to measure the computational complexity of the model. Table 2 shows the complexity of each component module. In the table, Head and Head small correspond to the input layer of the sample area UAV image and sample point UAV image, respectively; Resnet Blocks 1–3 are the sample area UAV image feature extraction layers; Resnet Block 2 small corresponds to the sample point UAV image feature extraction layer; Resnet Block 4 is the sample area, and sample point UAV image feature fusion. The Swin Transformer Block v2 is the function layer of feature fusion; Linear is the linear output layer.
The results in the table show that the overall Flops of ResTransformer are 1.99 × 109, and the total number of parameters is 1.20 × 107, while the Flops of the traditional Resnet50 model are around 2.91 × 109. The Flops of the ResTransformer model are 31.69% lower compared to Resnet50. Our proposed ResTransformer model has higher computational efficiency and lower complexity.

3.3. Hyperparameter Setting

The traditional Resnet residual network is good at solving the gradient disappearance and gradient explosion problems in deep network training. However, as its network layers increase, there is a degradation problem of network accuracy. To solve the problem that it is challenging to optimize the deep network, we use the Batch Normalization layer by adding it to accelerate the model convergence and improve the accuracy and generalization ability of the model. Furthermore, for the parameters in Resnet, we use a normal distribution for initialization, which achieves good accuracy on ImageNet. The Transformer architecture in the computer vision neighborhood is prone to model overfitting and bias due to the lack of structural bias and many parameters. To solve the problem of Swin Transformer Block v2 overfitting during training, we only use a single Swin Transformer Block v2 to solve the overfitting problem of Swin Transformer Block v2, we only use a single Swin Transformer Block v2 to calculate the output features of Resnet, and at the same time, we add a Dropout module to Swin Transformer Block v2 and use LayerNorm to normalize the output layer features in order to improve the generalization ability and robustness of the model.
During the overall training of the ResTransformer model, we generated enough sample data to improve the accuracy and generalization ability of the model. We used a 7:3 ratio to divide the training and test sets. A batch size of 32 was used during the training process; compared to a large batch size, training with a small batch size tends to converge to flat minimization regions that tend to generalize better because they are more robust to changes between the training and test sets [58]; epoch sets to 100, and the early stop strategy is activated, so that when the accuracy of the model on the test set has still not improved after three epochs, the training is stopped early. The model is considered to have completed optimization to reduce the risk of overfitting. The learning rate strategy is set to automatic decay learning rate strategy, the initial learning rate is 0.1, the focus round is set to 2, the decay factor is set to 0.8, and the cold start strategy is also configured. The variable learning rate can accelerate the convergence of the model; the optimizer adopts Adam optimizer, the weight decay of the optimizer is set to 0.0001, and the weight decay coefficient of L2 regularization is used to improve the convergence speed of the model. All input UAV images and DN values are divided by 255 to normalize their values to between [0, 1], and a fixed threshold is used to crop the gradient in the backpropagation process. The platform’s hardware, on which the training model is carried out in this paper, contains an Intel Xeon Gold 6244 CPU, 128 G RAM, and an Nvidia GeForce RTX 3090 series 24 G graphics card.

3.4. Evaluation Indicators

In this paper, the Simple Average method is employed as the baseline of the dataset, and Cubic Spline Interpolation and Kriging interpolation (Gaussian kernel function) methods are used on the dataset to verify the accuracy of various scale conversion methods, respectively. The results are compared with the results of the ResTransformer model. The mean relative error (MRE) between the DN value of the sample area and the mean DN value of the sample area UAV images obtained by each scale transformation method is calculated by Equation (4), where  D N p r e d  is the DN value of the sample area UAV grayscale images obtained by the scale transformation method,  D N t r u e  is the mean DN value of the sample area UAV grayscale images in the SCD dataset, and the root mean square error (RMSE) using Equation (5) is calculated to evaluate the accuracy of individual samples in the SCD dataset. The overall accuracy of each scale conversion method is evaluated by calculating the correlation coefficient  ρ  by Equation (6), where  C ov ( D N p r e d , D N t r u e )  is the covariance between the scale conversion result and the actual value in the data set, and  σ D N p r e d  and  σ D N t r u e  are the variances between the scale conversion result and the actual value, respectively. In summary, we verify the accuracy of various scale conversion methods using the above three methods.
M R E = i = 0 N D N p r e d ( i ) D N t r u e ( i ) D N t r u e ( i ) 100 %
R M S E = 1 N i = 1 N ( D N p r e d ( i ) D N t r u e ( i ) ) 2
ρ = C ov ( D N p r e d , D N t r u e ) σ D N p r e d σ D N t r u e

4. Results and Discussion

4.1. Accuracy Verification of Various Scale Conversion Methods

In this paper, the simple arithmetic averaging method, Cubic Spline Interpolation method, Ordinary Kriging square and ResTransformer deep learning model are used to validate the validation set of the SCD dataset, respectively. The results of the simple arithmetic averaging method are used as the baseline. Figure 9 shows the validation results, in which the scatter plot is plotted with the scale transformation results of each scale transformation model as the horizontal axis, the actual values of the SCD dataset as the vertical axis and the distribution are linearly fitted using the least squares method. The proper function of the conversion results for each scale shows that the slope of the fit function of the scale conversion result of the ResTransformer model is close to 1, and its scale conversion accuracy is the best; the slope of the fit function of the Ordinary Kriging method and that of the Simple Average method are both 0.72. The difference is less than 0.6%. Therefore, the results of the Ordinary Kriging method and the Simple Average method achieve similar accuracy on most of the data sets. The results of the Cubic Spline Interpolation method are more different from the actual values, and the distribution of the results is very discrete, so the method is less effective.
The MRE and RMSE of the scale conversion results of various scale conversion methods and the actual values of the SCD dataset were calculated separately to measure the accuracy of various scale conversion methods. Figure 10 shows the results, and the horizontal and vertical coordinates in the figure are the MRE and RMSE of the scale conversion results of various methods, respectively. The results show that the conversion accuracy of the Ordinary Kriging method is similar to that of the Simple Average method in most scenarios. The MRE and RMSE of most samples are concentrated in the ranges of 0.3–20% and 0.4–20%, respectively. The results of the Cubic Spline Interpolation method are much less accurate than the baseline of the Simple Average method in both MRE and RMSE metrics, and the distributions of MRE and RMSE metrics are more discrete. The ResTransformer deep learning model outperforms the traditional surface reflectance scale conversion method in MRE and RMSE, where the MRE results are less than 4%, and most of the RMSE results are less than 4. Therefore, the scale conversion accuracy of the ResTransformer model is better than the baseline of the Simple Average method.
Since there are a few samples with abnormally high MRE and RMSE metrics in the results of various scale conversion methods, there are some shortcomings in evaluating the accuracy of the corresponding scale conversion methods by using the mean values of MRE and RMSE of the scale conversion results on the SCD dataset, and Figure 11 shows the distribution of the accuracy of different scale conversion results. In order to evaluate the accuracy of various scale conversion methods as a whole, the accuracy of each scale conversion method was evaluated by calculating the median, the overall mean, and the mean value within the range of non-abnormal values (Equation (17), where  Q i  is the i × 25th percentile of the data), and the correlation coefficient between the scale conversion results and the actual values in the SCD dataset, respectively, of the MRE and RMSE metrics of the results of various scale conversion methods.
U p p e r = Q 3 + 1.5 ( Q 3 Q 1 ) L o w e r = Q 1 1.5 ( Q 3 Q 1 )
The results of the evaluation metrics calculated for each scale conversion method are shown in Table 3. The results show that among these four methods, the Simple Average and Ordinary Kriging methods have relatively high average MRE and average RMSE values of 8.56704 and 9.8842, and 8.60315 and 9.9272, respectively. This indicates that the mean error and root mean square error of these two methods are more considerable, i.e., they have lower prediction accuracy. However, both methods’ median MRE and median RMSE values are relatively low, indicating that they can yield relatively accurate predictions in some cases. The Cubic Spline Interpolation method is one of the worst performing methods with mean MRE and mean RMSE values of 49.77202 and 61.78559, respectively, which means that it has a very high mean error and root mean square error and is therefore not suitable for accurate prediction.
Moreover, it also has high IQR MRE and RMSE values, indicating that its predictions are less stable. In addition, its median MRE and median RMSE values are also high, indicating that the method generates significant prediction errors in some cases. In contrast, the mean MRE and RMSE values of the ResTransformer method are 0.6440 and 0.7460, respectively, which means that it has a low mean and mean square root error. At the same time, the ResTransformer model on Avg MRE, Avg IQR MRE, Avg RMSE, and Avg IQR RMSE all decreased by more than 90% compared to the baseline. The median decrease in MRE and RMSE was more significant than 85%, and the correlation coefficient improved by 16.59% compared to the baseline. The ResTransformer model had the best performance in all parameters relative to the baseline method. In addition, the ResTransformer method has the smallest IQR MRE and IQR RMSE values, indicating that the stability of its prediction result accuracy is also good. Overall, ResTransformer is the best method. Simple Average and Ordinary Kriging can provide better prediction results in some cases, while the Cubic Spline Interpolation method performs the worst.

4.2. Effect of the Number of Sampling Points and Sample Area on the Accuracy of Scale Conversion Results

In surface reflectance measurement, the accuracy of surface reflectance at the image scale will be changed when different sampling methods are used for the same size sample area and the same sampling methods are used for different size sample areas. The average MREs of the traditional medium- and high-resolution satellite image size samples were calculated under different standard sampling methods, including 30 m, 20 m, 10 m, 8 m, 4 m and 2 m, corresponding to Landsat, Spot, Sentinel and GF series satellites. The standard sampling methods include one-point, two-point, four-point, five-point, nine-point, and sixteen-point sampling methods.
Figure 12 shows the scale conversion results’ accuracy with different sampling points and sample areas. The figure shows that the scale conversion results of the Simple Average method and ResTransformer model, in the same size of a sample area, the MRE gradually decreases as the number of sampling points increases. When the sampling method is fixed, the MRE does not strictly increase with the increase of sample area, and the MRE is in a fluctuating rising state. We found that the error of the same scale conversion method in the 4 m sample area is much higher than that in the 20 m sample area under the sampling method with a smaller number of sampling points. Therefore, the number of sampling points is the main factor affecting the accuracy of the scale conversion results in the actual sampling process.
Table 4 and Table 5 quantify the influence of the number of sampling points and sample area size on the accuracy of the scale conversion results by calculating the correlation coefficients between the number of sampling points and the MRE of the scale conversion results under the same size sample area and the correlation coefficients between the sample area size and the MRE of the scale conversion results under the same number of sampling points. The results revealed that the MRE of the scale conversion results showed an apparent negative correlation with the number of sampling points. In contrast, the correlation with the sample area was not significant, especially when the fixed number of sampling points was 2 and 4, and there was no linear correlation between the MRE of the scale conversion results and the sample area size. Therefore, in the actual sampling process, the number of sampling points was the main factor affecting the accuracy of the scale conversion results.
In the field image-scale geo-spectral acquisition activities, if we need to carry out star–earth synchronous matching calibration of various satellite sensors through the measured image scale geo-spectra, then we need to increase the time cost, reduce the efficiency, and increase the number of sampling points as much as possible to obtain the image scale geo-spectral data with higher accuracy. Field image-scale geo-spectral acquisition also includes automatic instrument acquisition, which usually involves installing automatic geo-spectral acquisition devices at fixed locations. Since it is costly to add automatic acquisition instruments, the number of instruments is fixed. The accuracy of image scale geo-spectral can only be improved by improving scale conversion methods.
During routine field spectral acquisition, too many sampling points will lead to an increase in the chance of the sample area being destroyed by human factors and an increase in the uncertainty caused by human factors. This means that the accuracy of the scale conversion results cannot be improved by increasing the number of sampling points, so it is necessary to achieve the least number of sampling methods while ensuring a certain accuracy. As can be seen from Figure 12, when the number of sampling points is increased from 1 to 4, the ResTransformer scale conversion method results in an average decrease of 35.3% in MRE and a 4-fold increase in workload, and when the number is increased from 1 to 16, the scale conversion results in an average decrease of 45.3% in MRE, but a 16-fold increase in sampling workload. Similarly, when the number of sampling points is increased from 1 to 4, the MRE of the Simple Average scale conversion method results in an average decrease of 33%, and when the number is increased from 1 to 16, the MRE of the scale conversion results in an average decrease of 74.3%. As the number of samples increases, the rate of MRE decreases, slows down, and the sampling efficiency decreases. Therefore, when the accuracy of four-point sampling meets the required conditions, increasing the number of additional sampling points becomes unnecessary.

4.3. Effect of Different Sub-Bedding Surfaces on the Accuracy of Scale Conversion Results

In the process of surface reflectance field collection, the accuracy of the scale conversion results will change due to the different heterogeneity of different feature types. To investigate the influence of different feature types on the scale conversion accuracy, the Simple Average and ResTransformer methods were used to scale convert 34 different feature types in the SCD dataset, and the average MRE values of the scale conversion results of different feature types were calculated.
Figure 13 shows the average accuracy of the scale conversion results for different subsurface samples. The figure shows that the Simple Average method and the ResTransformer model have similar trends in the accuracy of the scale conversion results for different feature types. Meanwhile, the figure shows that the accuracy of the scale conversion results obtained by different scale conversion methods is higher for the feature types with homogeneous ground surfaces and low ground-cover height. The typical features of this type include non-vegetated ground surfaces such as water bodies, asphalt, concrete, bare soil and sand, and intensive vegetation covers such as grass and straw. For the type of surface cover with higher vegetation and dense vegetation, rarely seen bare soil between the monopoly, the accuracy of the scale conversion results of different scale conversion methods is higher compared with the types of sand and water bodies. However, the number of sampling points can be increased to improve sure accuracy, and the typical features of this type include dense vegetation cover such as cotton, sweet potato, okra, peanut and soybean. For monopoly row vegetation, higher crops and sparse, non-uniformly distributed ground types, the accuracy of the scale conversion results is lower. Typical ground types include corn, weeds and vegetation such as red-fruited canary.
To quantitatively describe the influence of spatial heterogeneity on the accuracy of scale conversion results of various scale conversion methods, the coefficients of variation of different samples were calculated by Equation (18), where  s t d D N  and  m e a n D N  are the standard deviation and mean of DN values of UAV grayscale images in the sample area, respectively. The correlation between the coefficients of variation and the accuracy of scale conversion results was checked to quantitatively describe the influence of spatial heterogeneity on the accuracy of scale conversion results of various scale conversion methods.
c v = s t d D N m e a n D N 100 %
Figure 14 shows the correlation between the surface heterogeneity and the accuracy of the scale transformation. The results in the figure show that the linear relationship between the MRE of the Simple Average method and the ResTransformer model results and the coefficient of variation of the sample area is more significant, and the R2 of the linear fit is 0.75, indicating that the accuracy of the scale transformation method results decreases gradually with the increase of the surface heterogeneity. The R2 of the linear fit is 0.75, indicating that the accuracy of the scale conversion method decreases as the surface heterogeneity increases. Meanwhile, the MRE of the Simple Average method changes rapidly with the increase in coefficient of variation of the sample area, and the slope of the fitted straight line is 0.21. The MRE of the ResTransformer model changes slowly with the increase in coefficient of variation of the sample area, proving that the ResTransformer model has higher robustness and accuracy than the Simple Average method. It is proved that the ResTransformer model has higher robustness and universality compared with the Simple Average method and can adapt to scale transformation tasks in more complex sub-bedding scenarios.

5. Conclusions

In this paper, we convert the scale conversion problem of surface reflectance to the scale conversion problem of grayscale values of sampling points in UAV images by imputation and prove the correctness of imputation. This is achieved by establishing a functional relationship between the two, and employing the ResTransformer deep learning model to extract and fuse features. The ResTransformer deep learning model is also established to extract, fuse, and adaptively learn the features within the UAV image of the target sample area and the UAV image of the sample points within the sample area, and determine the high-dimensional nonlinear spatial relationship between the sample points and the sample area within the target sample area. This enables realization of the scale conversion of the surface reflectance at the image element scale with high accuracy under the scenarios of arbitrary size, heterogeneous sample area and sampling mode. To verify the accuracy and robustness of ResTransformer, a scale-transformed dataset including 500k samples is built in this paper. Moreover, after being validated through various traditional methods with ResTransformer, the results show that the ResTransformer deep learning model scales the transformation results on the SCD dataset with much better accuracy than the Cubic Spline Interpolation and Ordinary Kriging methods, while compared to the Simple Average method, the baseline is greatly improved. The mean MRE, mean MRSE, and correlation coefficient R with the actual value for ResTransformer on the SCD dataset were 0.6440%, 0.7460, and 0.99911, respectively.
In comparison, the accuracy of the baseline on the SCD dataset was 8.5670%, 9.8842, and 0.8569, respectively. The improvements of ResTransformer accuracy compared to the baseline were 92.48%, 92.45% and 16.59%, respectively. Compared with the baseline method, ResTransformer offers high robustness and is suitable for different sampling methods and sample areas of different sizes. In contrast, the accuracy of the baseline method changes significantly with a decrease in the number of sampling points and an increase in the sample area. Hence, the method is only suitable in certain applications. In future research, we plan to iteratively update the model by designing and building a large model as well as creating a large sample dataset with big data. Further, we plan to extend the imputation method, as well as the model, from surface reflectance to other surface quantitative remote sensing parameters, such as surface temperature and LAI.

Author Contributions

Conceptualization, funding acquisition, writing—review and editing, project administration, H.G.; Conceptualization, methodology, writing—original draft preparation, software, validation, X.Q.; investigation, Y.W.; resources, F.L., Y.Y., Z.L. and W.Z.; data curation, X.S.; visualization, supervision, Q.W.; formal analysis, X.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Strategic Priority Research Program of the Chinese Academy of Sciences (XDA28050401), China’s 13th Five-Year Plan Civil Space Pre-Research Project (EOA203010F) and Ecological environment satellite star-ground synchronous authenticity verification experiment (E2C2031501).

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. DeCoffe, L.J.R.; Conran, D.N.; Bauch, T.D.; Ross, M.G.; Kaputa, D.S.; Salvaggio, C. Initial Performance Analysis of the At-Altitude Radiance Ratio Method for Reflectance Conversion of Hyperspectral Remote Sensing Data. Sensors 2023, 23, 320. [Google Scholar] [CrossRef] [PubMed]
  2. Hao, D.; Xiao, Q.; Wen, J.; You, D.; Wu, X.; Lin, X.; Wu, S. Advances in upscaling methods of quantitative remote sensing. J. Remote Sens. 2018, 22, 408–423. [Google Scholar]
  3. Meng, B.; Wang, J. A Review on the Methodology of Scaling with Geo-Data. Acta Geogr. Sin. 2005, 60, 277–288. [Google Scholar]
  4. Luan, H.; Tian, Q.; Yu, T.; Hu, X.; Huang, Y.; Liu, L.; Du, L.; Wei, X. Review of Up-scaling of Quantitative Remote Sensing. Adv. Earth Sci. 2013, 28, 657–664. [Google Scholar]
  5. Wu, X.; Wen, J.; Xiao, Q.; Li, X.; Liu, Q.; Tang, Y.; Dou, B.; Peng, J.; You, D.; Li, X. Advances in validation methods for remote sensing products of land surface parameters. J. Remote Sens. 2015, 19, 75–92. [Google Scholar]
  6. Yu, Y.; Pan, Y.; Yang, X.G.; Fan, W.Y. Spatial Scale Effect and Correction of Forest Aboveground Biomass Estimation Using Remote Sensing. Remote Sens. 2022, 14, 2828. [Google Scholar] [CrossRef]
  7. Naethe, P.; Asgari, M.; Kneer, C.; Knieps, M.; Jenal, A.; Weber, I.; Moelter, T.; Dzunic, F.; Deffert, P.; Rommel, E.; et al. Calibration and Validation from Ground to Airborne and Satellite Level: Joint Application of Time-Synchronous Field Spectroscopy, Drone, Aircraft and Sentinel-2 Imaging. Pfg-J. Photogramm. Remote Sens. Geoinf. Sci. 2023, 91, 43–58. [Google Scholar] [CrossRef]
  8. Tang, H.Z.; Xie, J.F.; Chen, W.; Zhang, H.G.; Wang, H.Y. Absolute Radiometric Calibration of ZY3-02 Satellite Multispectral Imager Based on Irradiance-Based Method. Remote Sens. 2023, 15, 448. [Google Scholar] [CrossRef]
  9. Tang, H.Z.; Xie, J.F.; Tang, X.M.; Chen, W.; Li, Q. On-Orbit Radiometric Performance of GF-7 Satellite Multispectral Imagery. Remote Sens. 2022, 14, 886. [Google Scholar] [CrossRef]
  10. Thome, K.J. Absolute radiometric calibration of Landsat 7 ETM+ using the reflectance-based method. Remote Sens. Environ. 2001, 78, 27–38. [Google Scholar] [CrossRef]
  11. Hufkens, K.; Bogaert, J.; Dong, Q.H.; Lu, L.; Huang, C.L.; Ma, M.G.; Che, T.; Li, X.; Veroustraete, F.; Ceulemans, R. Impacts and uncertainties of upscaling of remote-sensing data validation for a semi-arid woodland. J. Arid. Environ. 2008, 72, 1490–1505. [Google Scholar] [CrossRef]
  12. Shi, Y.; Wang, J.; Qin, J.; Qu, Y. An Upscaling Algorithm to Obtain the Representative Ground Truth of LAI Time Series in Heterogeneous Land Surface. Remote Sens. 2015, 7, 12887–12908. [Google Scholar] [CrossRef]
  13. Wang, K.C.; Liu, J.M.; Zhou, X.J.; Sparrow, M.; Ma, M.; Sun, Z.; Jiang, W.H. Validation of the MODIS global land surface albedo product using ground measurements in a semidesert region on the Tibetan Plateau. J. Geophys. Res. Atmos. 2004, 109. [Google Scholar] [CrossRef]
  14. Li, X.; Jin, R.; Liu, S.; Ge, Y.; Xiao, Q.; Liu, Q.; Ma, M.; Ran, Y. Upscaling research in HiWATER: Progress and prospects. J. Remote Sens. 2016, 20, 921–932. [Google Scholar]
  15. Crow, W.T.; Ryu, D.; Famiglietti, J.S. Upscaling of field-scale soil moisture measurements using distributed land surface modeling. Adv. Water Resour. 2005, 28, 1–14. [Google Scholar] [CrossRef]
  16. Qin, J.; Yang, K.; Lu, N.; Chen, Y.; Zhao, L.; Han, M. Spatial upscaling of in-situ soil moisture measurements based on MODIS-derived apparent thermal inertia. Remote Sens. Environ. 2013, 138, 1–9. [Google Scholar] [CrossRef]
  17. Wu, X.; Wen, J.; Xiao, Q.; Liu, Q.; Peng, J.; Dou, B.; Li, X.; You, D.; Tang, Y.; Liu, Q. Coarse scale in situ albedo observations over heterogeneous snow-free land surfaces and validation strategy: A case of MODIS albedo products preliminary validation over northern China. Remote Sens. Environ. 2016, 184, 25–39. [Google Scholar] [CrossRef]
  18. Li, X. Characterization, controlling, and reduction of uncertainties in the modeling and observation of land-surface systems. Sci. China-Earth Sci. 2014, 57, 80–87. [Google Scholar] [CrossRef]
  19. Erickson, T.A.; Williams, M.W.; Winstral, A. Persistence of topographic controls on the spatial distribution of snow in rugged mountain terrain, Colorado, United States. Water Resour. Res. 2005, 41. [Google Scholar] [CrossRef]
  20. Hengl, T.; Heuvelink, G.B.M.; Rossiter, D.G. About regression-kriging: From equations to case studies. Comput. Geosci. 2007, 33, 1301–1315. [Google Scholar] [CrossRef]
  21. Wen, J.; Liu, Q.; Liu, Q.; Xiao, Q.; Li, X. Scale effect and scale correction of land-surface albedo in rugged terrain. Int. J. Remote Sens. 2009, 30, 5397–5420. [Google Scholar] [CrossRef]
  22. Christakos, G.; Kolovos, A.; Serre, M.L.; Vukovich, F. Total ozone mapping by integrating databases from remote sensing instruments and empirical models. IEEE Trans. Geosci. Remote Sens. 2004, 42, 991–1008. [Google Scholar] [CrossRef]
  23. Gao, S.; Zhu, Z.; Liu, S.; Jin, R.; Yang, G.; Tan, L. Estimating the spatial distribution of soil moisture based on Bayesian maximum entropy method with auxiliary data from remote sensing. Int. J. Appl. Earth Obs. Geoinf. 2014, 32, 54–66. [Google Scholar] [CrossRef]
  24. Du, J.; Zhou, H.H.; Jacinthe, P.A.; Song, K.S. Retrieval of lake water surface albedo from Sentinel-2 remote sensing imagery. J. Hydrol. 2023, 617, 128904. [Google Scholar] [CrossRef]
  25. Liang, S.L.; Fang, H.L.; Chen, M.Z.; Shuey, C.J.; Walthall, C.; Daughtry, C.; Morisette, J.; Schaaf, C.; Strahler, A. Validating MODIS land surface reflectance and albedo products: Methods and preliminary results. Remote Sens. Environ. 2002, 83, 149–162. [Google Scholar] [CrossRef]
  26. Yue, X.Y.; Li, Z.Q.; Li, H.L.; Wang, F.T.; Jin, S. Multi-Temporal Variations in Surface Albedo on Urumqi Glacier No.1 in Tien Shan, under Arid and Semi-Arid Environment. Remote Sens. 2022, 14, 808. [Google Scholar] [CrossRef]
  27. Sanchez-Zapero, J.; Martinez-Sanchez, E.; Camacho, F.; Wang, Z.S.; Carrer, D.; Schaaf, C.; Garcfa-Haro, F.J.; Nickeson, J.; Cosh, M. Surface ALbedo VALidation (SALVAL) Platform: Towards CEOS LPV Validation Stage 4-Application to Three Global Albedo Climate Data Records. Remote Sens. 2023, 15, 1081. [Google Scholar] [CrossRef]
  28. Jin, R.; Li, X.; Liu, S.M. Understanding the Heterogeneity of Soil Moisture and Evapotranspiration Using Multiscale Observations From Satellites, Airborne Sensors, and a Ground-Based Observation Matrix. IEEE Geosci. Remote Sens. Lett. 2017, 14, 2132–2136. [Google Scholar] [CrossRef]
  29. Bachmann, C.M.; Montes, M.J.; Parrish, C.E.; Fusina, R.A.; Nichols, C.R.; Li, R.R.; Hallenborg, E.; Jones, C.A.; Lee, K.; Sellars, J.; et al. A dual-spectrometer approach to reflectance measurements under sub-optimal sky conditions. Opt. Express 2012, 20, 8959–8973. [Google Scholar] [CrossRef]
  30. Francos, N.; Ben-Dor, E. A transfer function to predict soil surface reflectance from laboratory soil spectral libraries. Geoderma 2022, 405, 115432. [Google Scholar] [CrossRef]
  31. Lee, Z.P.; Ahn, Y.H.; Mobley, C.; Arnone, R. Removal of surface-reflected light for the measurement of remote-sensing reflectance from an above-surface platform. Opt. Express 2010, 18, 26313–26324. [Google Scholar] [CrossRef]
  32. Glasser, C.; Selman, A.L.; Sengupta, S. Reductions between disjoint NP-pairs. Inf. Comput. 2005, 200, 247–267. [Google Scholar] [CrossRef]
  33. Buhrman, H.; Fortnow, L.; Hitchcock, J.M.; Loff, B. Learning Reductions to Sparse Sets. In Proceedings of the 38th International Symposium on Mathematical Foundations of Computer Science (MFCS), IST Austria, Klosterneuburg, Austria, 26–30 August 2013; pp. 243–253. [Google Scholar]
  34. Hitchcock, J.M.; Shafei, H. Nonuniform Reductions and NP-Completeness. Theory Comput. Syst. 2022, 66, 743–757. [Google Scholar] [CrossRef]
  35. Campelo, M.; Campos, V.A.; Correa, R.C. On the asymmetric representatives formulation for the vertex coloring problem. Discret. Appl. Math. 2008, 156, 1097–1111. [Google Scholar] [CrossRef]
  36. Zhu, E.Q.; Jiang, F.; Liu, C.J.; Xu, J. Partition Independent Set and Reduction-Based Approach for Partition Coloring Problem. IEEE Trans. Cybern. 2022, 52, 4960–4969. [Google Scholar] [CrossRef] [PubMed]
  37. Cao, C.; Lee, X.; Muhlhausen, J.; Bonneau, L.; Xu, J.P. Measuring Landscape Albedo Using Unmanned Aerial Vehicles. Remote Sens. 2018, 10, 1812. [Google Scholar] [CrossRef]
  38. Duan, S.B.; Li, Z.L.; Tang, B.H.; Wu, H.; Ma, L.L.; Zhao, E.Y.; Li, C.R. Land Surface Reflectance Retrieval from Hyperspectral Data Collected by an Unmanned Aerial Vehicle over the Baotou Test Site. PLoS ONE 2013, 8, e66972. [Google Scholar] [CrossRef]
  39. Wang, R.; Gamon, J.A.; Hmimina, G.; Cogliati, S.; Zygielbaum, A.I.; Arkebauer, T.J.; Suyker, A. Harmonizing solar induced fluorescence across spatial scales, instruments, and extraction methods using proximal and airborne remote sensing: A multi-scale study in a soybean field. Remote Sens. Environ. 2022, 281, 113268. [Google Scholar] [CrossRef]
  40. Helder, D.; Thome, K.; Aaron, D.; Leigh, L.; Czapla-Myers, J.; Leisso, N.; Biggar, S.; Anderson, N. Recent surface reflectance measurement campaigns with emphasis on best practices, SI traceability and uncertainty estimation. Metrologia 2012, 49, S21–S28. [Google Scholar] [CrossRef]
  41. Walsh, A.J.; Byrne, G.; Broomhall, M. A Case Study of Measurement Uncertainty in Field Spectroscopy. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 6248–6258. [Google Scholar] [CrossRef]
  42. Krichen, M.; Mihoub, A.; Alzahrani, M.Y.; Adoni, W.Y.H.; Nahhal, T. Are Formal Methods Applicable To Machine Learning And Artificial Intelligence? In Proceedings of the 2022 2nd International Conference of Smart Systems and Emerging Technologies (SMARTTECH), Riyadh, Saudi Arabia, 9–11 May 2022; pp. 48–53. [Google Scholar]
  43. Seshia, S.A.; Sadigh, D.; Sastry, S.S. Toward verified artificial intelligence. Commun. ACM 2022, 65, 46–55. [Google Scholar] [CrossRef]
  44. Sun, H.; Pan, J.; Gao, H.L.; Jiang, P.; Dou, X.G.; Wang, K.S.; Chen, L.; Li, L.; Han, Q.J. Sampling Method and Accuracy of Pixel-Scale Surface Reflectance at Dunhuang Site. Laser Optoelectron. Prog. 2022, 59, 1028009. [Google Scholar] [CrossRef]
  45. Hengl, T.; Heuvelink, G.B.M.; Stein, A. A generic framework for spatial prediction of soil variables based on regression-kriging. Geoderma 2004, 120, 75–93. [Google Scholar] [CrossRef]
  46. Oliver, M.A.; Webster, R. A tutorial guide to geostatistics: Computing and modelling variograms and kriging. Catena 2014, 113, 56–69. [Google Scholar] [CrossRef]
  47. Pebesma, E.J. Multivariable geostatistics in S: The gstat package. Comput. Geosci. 2004, 30, 683–691. [Google Scholar] [CrossRef]
  48. Bogdanov, V.V.; Volkov, Y.S. Near-optimal tension parameters in convexity preserving interpolation by generalized cubic splines. Numer. Algorithms 2021, 86, 833–861. [Google Scholar] [CrossRef]
  49. Dai, X.; Wang, G.; Wang, P. Wavelet Sampling and Meteorological Record Interpolation. Chin. J. Comput. Phys. 2003, 20, 529–536. [Google Scholar]
  50. Duan, Q.; Djidjeli, K.; Price, W.G.; Twizell, E.H. Weighted rational cubic spline interpolation and its application. J. Comput. Appl. Math. 2000, 117, 121–135. [Google Scholar] [CrossRef]
  51. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  52. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention Is All You Need. In Proceedings of the 31st Annual Conference on Neural Information Processing Systems (NIPS), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  53. Devlin, J.; Chang, M.-W.; Lee, K.; Toutanova, K.; Assoc Computat, L. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the North-American-Chapter of the Association-for-Computational-Linguistics—Human Language Technologies (NAACL-HLT), Minneapolis, MN, USA, 2–7 June 2019; pp. 4171–4186. [Google Scholar]
  54. Raffel, C.; Shazeer, N.; Roberts, A.; Lee, K.; Narang, S.; Matena, M.; Zhou, Y.; Li, W.; Liu, P.J. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer. J. Mach. Learn. Res. 2020, 21, 5485–5551. [Google Scholar]
  55. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  56. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer using Shifted Windows. In Proceedings of the 18th IEEE/CVF International Conference on Computer Vision (ICCV), Electr Network, Online, 11–17 October 2021; pp. 9992–10002. [Google Scholar]
  57. Liu, Z.; Hu, H.; Lin, Y.; Yao, Z.; Xie, Z.; Wei, Y.; Ning, J.; Cao, Y.; Zhang, Z.; Dong, L.; et al. Swin Transformer V2: Scaling Up Capacity and Resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 11999–12009. [Google Scholar]
  58. Nocedal, J.; Keskar, N.; Mudigere, D.; Tang, P.; Smelyanskiy, M. Large-Batch Training for Deep Learning: Generalization Gap and Sharp Minima. arXiv 2017, arXiv:1609.04836. [Google Scholar]
Figure 1. Schematic diagram of surface-reflectance scale conversion reduction. A,B,C,D are the numbers of sampling points respectively.
Figure 1. Schematic diagram of surface-reflectance scale conversion reduction. A,B,C,D are the numbers of sampling points respectively.
Remotesensing 15 02449 g001
Figure 2. UAV image acquisition location and feature type.
Figure 2. UAV image acquisition location and feature type.
Remotesensing 15 02449 g002
Figure 3. Overall process of SCD dataset construction.
Figure 3. Overall process of SCD dataset construction.
Remotesensing 15 02449 g003
Figure 4. UAV image sample-enhancement scheme.
Figure 4. UAV image sample-enhancement scheme.
Remotesensing 15 02449 g004
Figure 5. Schematic illustration of the location of sampling points for various sampling methods.
Figure 5. Schematic illustration of the location of sampling points for various sampling methods.
Remotesensing 15 02449 g005
Figure 6. Construction process for a single sample of the SCD dataset.
Figure 6. Construction process for a single sample of the SCD dataset.
Remotesensing 15 02449 g006
Figure 7. SCD dataset sample encoding method.
Figure 7. SCD dataset sample encoding method.
Remotesensing 15 02449 g007
Figure 8. ResTransformer model structure.
Figure 8. ResTransformer model structure.
Remotesensing 15 02449 g008
Figure 9. Scale conversion results for multiple scale conversion methods (a) Simple Average method, (b) Cubic Spline Interpolation method, (c) Ordinary Kriging method, (d) ResTransformer model.
Figure 9. Scale conversion results for multiple scale conversion methods (a) Simple Average method, (b) Cubic Spline Interpolation method, (c) Ordinary Kriging method, (d) ResTransformer model.
Remotesensing 15 02449 g009
Figure 10. Accuracy of scale conversion results of multiple scale conversion methods (a) Simple Average method, (b) Cubic Spline Interpolation method, (c) Ordinary Kriging method, (d) ResTransformer model.
Figure 10. Accuracy of scale conversion results of multiple scale conversion methods (a) Simple Average method, (b) Cubic Spline Interpolation method, (c) Ordinary Kriging method, (d) ResTransformer model.
Remotesensing 15 02449 g010
Figure 11. Distribution of MRE and RMSE as a result of various scale conversion methods. The solid dots in the graph indicate outliers exceeding 1.5 times the IQR, and the hollow circles indicate the average values.
Figure 11. Distribution of MRE and RMSE as a result of various scale conversion methods. The solid dots in the graph indicate outliers exceeding 1.5 times the IQR, and the hollow circles indicate the average values.
Remotesensing 15 02449 g011
Figure 12. Accuracy of scale transformation results with different numbers of sampling points and sample area. (a) Simple Average method. (b) ResTransformer model.
Figure 12. Accuracy of scale transformation results with different numbers of sampling points and sample area. (a) Simple Average method. (b) ResTransformer model.
Remotesensing 15 02449 g012
Figure 13. Average precision of the conversion results of sample scales for different feature types. The red, green, and blue bars in the figure represent the types of surface coverage with high, medium, and low accuracy for surface reflectance scale conversion, respectively.
Figure 13. Average precision of the conversion results of sample scales for different feature types. The red, green, and blue bars in the figure represent the types of surface coverage with high, medium, and low accuracy for surface reflectance scale conversion, respectively.
Remotesensing 15 02449 g013
Figure 14. Correlation between the coefficient of variation of the sample area and the MRE of the scale conversion results of various scale conversion methods.
Figure 14. Correlation between the coefficient of variation of the sample area and the MRE of the scale conversion results of various scale conversion methods.
Remotesensing 15 02449 g014
Table 1. Number of UAV images of different feature types and corresponding observation areas.
Table 1. Number of UAV images of different feature types and corresponding observation areas.
Feature TypeNumArea km2Feature TypeNumArea km2Feature TypeNumArea km2
Chili561.83 × 100Corn2116.09 × 100Sweet potato362.16 × 100
Rice1431.52 × 101Eggplant63.08× 101Orange trees421.10 × 100
Wheat521.56 × 101Cauliflower481.26 × 100Water bodies733.69 × 100
Grassland1364.92 × 100Peanut112.00× 101Chrysanthemum208.02× 101
Cotton83.01 × 101Soybean153.21× 101green cabbage122.35 × 100
Zucchini743.33 × 100Sandy568.36 × 100Lover’s Grass881.96 × 100
Concrete657.38 × 100Tomatoes2226.14 × 100Fluffy grass1854.87 × 100
Greenhouse864.94 × 100Bare soil4823.39× 101Golden peach603.63 × 100
Okra53.40 × 101Date palm23.53× 101Purple kale1112.89 × 100
Pitch161.14 × 101Open space22.68× 101Green onions1325.98 × 100
Beans142.82 × 100Gravel23.15× 101
Weeds127.87 × 102Straw472.77 × 100
Table 2. Complexity of each module of the ResTransformer model.
Table 2. Complexity of each module of the ResTransformer model.
NameInput ShapeOutput ShapeFlopsFlops PercentageParams
Head[3, 224, 224][64, 128, 128]1.20 × 1086.008%9.54 × 103
Head Small[48, 56, 56][64, 28, 28]1.19 × 1085.973%1.51 × 105
Resnet Block 1[64, 128, 128][64, 56, 56]4.64 × 10823.327%1.48 × 105
Resnet Block 2[64, 56, 56][128, 28, 28]4.12 × 10820.706%5.26 × 105
Resnet Block 3[128, 28, 28][256, 14, 14]4.12 × 10820.676%2.10 × 106
Resnet Block 2 Small[64, 28, 28][64, 14, 14]4.27 × 10821.467%8.72 × 106
Resnet Block 4[320, 14, 14][512, 7, 7]2.99 × 1071.500%1.52 × 105
Swin Transformer Block v2[49, 256][49, 256]6.81 × 1060.342%1.51 × 105
Linear[49, 256][1]1.75 × 1040.001%1.75 × 104
TOTAL--1.99 × 109100%1.20 × 107
Table 3. Average accuracy of various scale conversion methods on the SCD dataset.
Table 3. Average accuracy of various scale conversion methods on the SCD dataset.
Method NameAvg MRE (%)Avg RMSEAvg IQR MRE (%)Avg IQR RMSEMedian MRE (%)Median RMSER
Simple Average8.567049.88426.169477.432614.587995.674520.85693
Cubic Spline Interpolation49.7720261.7855931.2214439.9374428.4603831.06510.40127
Ordinary Kriging8.603159.92726.200617.490484.627845.719120.85527
ResTransformer0.64400.74600.523350.622970.64400.54900.99911
Table 4. Correlation between the number of sampling points and the scale conversion result MRE under the same size sample area.
Table 4. Correlation between the number of sampling points and the scale conversion result MRE under the same size sample area.
Scale Conversion MethodSample Area Edge Length (m)RScale Conversion MethodSample Area Edge Length (m)R
Simple Average2−0.8618ResTransformer2−0.7048
4−0.84314−0.7299
8−0.84558−0.7987
10−0.821110−0.7461
20−0.862620−0.7371
30−0.852630−0.7627
Table 5. The correlation between sample area size and scale conversion results MRE under the same number of sampling points.
Table 5. The correlation between sample area size and scale conversion results MRE under the same number of sampling points.
Scale Conversion MethodNumber of Sampling PointsRScale Conversion MethodNumber of Sampling PointsR
Simple Average1−0.4412ResTransformer1−0.6118
2−0.000820.0346
4−0.170340.0061
5−0.151550.5472
90.320490.5243
160.5300160.5608
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Qiu, X.; Gao, H.; Wang, Y.; Zhang, W.; Shi, X.; Lv, F.; Yu, Y.; Luan, Z.; Wang, Q.; Zhao, X. A Scale Conversion Model Based on Deep Learning of UAV Images. Remote Sens. 2023, 15, 2449. https://doi.org/10.3390/rs15092449

AMA Style

Qiu X, Gao H, Wang Y, Zhang W, Shi X, Lv F, Yu Y, Luan Z, Wang Q, Zhao X. A Scale Conversion Model Based on Deep Learning of UAV Images. Remote Sensing. 2023; 15(9):2449. https://doi.org/10.3390/rs15092449

Chicago/Turabian Style

Qiu, Xingchen, Hailiang Gao, Yixue Wang, Wei Zhang, Xinda Shi, Fengjun Lv, Yanqiu Yu, Zhuoran Luan, Qianqian Wang, and Xiaofei Zhao. 2023. "A Scale Conversion Model Based on Deep Learning of UAV Images" Remote Sensing 15, no. 9: 2449. https://doi.org/10.3390/rs15092449

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop