The Impact of Spatial Resolution on the Classification of Vegetation Types in Highly Fragmented Planting Areas Based on Unmanned Aerial Vehicle Hyperspectral Images

Fine classification of vegetation types has always been the focus and difficulty in the application field of remote sensing. Unmanned Aerial Vehicle (UAV) sensors and platforms have become important data sources in various application fields due to their high spatial resolution and flexibility. Especially, UAV hyperspectral images can play a significant role in the fine classification of vegetation types. However, it is not clear how the ultrahigh resolution UAV hyperspectral images react in the fine classification of vegetation types in highly fragmented planting areas, and how the spatial resolution variation of UAV images will affect the classification accuracy. Based on UAV hyperspectral images obtained from a commercial hyperspectral imaging sensor (S185) onboard a UAV platform, this paper examines the impact of spatial resolution on the classification of vegetation types in highly fragmented planting areas in southern China by aggregating 0.025 m hyperspectral image to relatively coarse spatial resolutions (0.05, 0.1, 0.25, 0.5, 1, 2.5 m). The object-based image analysis (OBIA) method was used and the effects of several segmentation scale parameters and different number of features were discussed. Finally, the classification accuracies from 84.3% to 91.3% were obtained successfully for multi-scale images. The results show that with the decrease of spatial resolution, the classification accuracies show a stable and slight fluctuation and then gradually decrease since the 0.5 m spatial resolution. The best classification accuracy does not occur in the original image, but at an intermediate level of resolution. The study also proves that the appropriate feature parameters vary at different scales. With the decrease of spatial resolution, the importance of vegetation index features has increased, and that of textural features shows an opposite trend; the appropriate segmentation scale has gradually decreased, and the appropriate number of features is 30 to 40. Therefore, it is of vital importance to select appropriate feature parameters for images in different scales so as to ensure the accuracy of classification.


Introduction
Vegetation is the main component of the ecosystem and plays an important role in the process of material circulation and energy exchange on land surface. Various vegetation types have different responses to the ecosystem. Monitoring vegetation types is of great significance for mastering its current shows that the identification accuracy of 2.4 m images is the highest [31]. Roth spatially aggregated the fine resolution (3-18 m) of airborne AVIRIS to the coarser resolution (20-60 m) for accurate mapping of plant species. The result shows that the best classification accuracy is at the coarser resolution, not the original image [32]. Despite these achievements, most research has been based on images at meter-level and sub-meter-level resolutions. However, for more fragmented planting structures and based on finer spatial resolution (e.g., centimeter level), how to balance the relationship between monitoring target scale and image scale and on which resolution the vegetation types can be accurately identified have to be solved. Especially in view of the limited ground coverage of UAV images, it is of great significance to flight at a proper spatial scale to maximize the coverage while ensuring classification accuracy.
This study is mainly aimed at: (1) Making full use of the hyperspectral and high spatial characteristics of UAV hyperspectral images to realize the fine classification of vegetation types in highly fragmented planting areas and (2) obtaining the scale variation characteristics of vegetation type identification for UAV images, exploring the appropriate scale range to provide reference for UAV flight experiment design, remote sensing image selection and UAV image application of vegetation classification in similar areas.

Overview of the Study Area
The study area is located in Xingbin District, Laibin City, Guangxi, China, East Asia ( Figure 1). Laibin City is located between 108 • 24 -110 • 28 E and 23 • 16 -24 • 29 N, with mountains and hills accounting for 65%, and other areas are platforms and plains. It boasts a subtropical monsoon climate with warm weather, abundant rain and insignificant seasonal changes. Laibin City is rich in vegetation resources, which has the typical characteristics of complex planting structure, abnormal fragmented field parcels, terrain cutting in southern China. Different vegetation types show irregular and staggered distribution, which brings great difficulties to the fine classification of vegetation types. The main vegetation types in Laibin City are sugarcane, eucalyptus, citrus, etc. Laibin City is one of the key sugarcane production bases in China, taking the sugarcane production as the pillar industry. It has obvious advantages in planting fast-growing forests, and high-quality eucalyptus trees are widely planted. In addition, there are many kinds of horticultural crops in Laibin City, of which citrus is the most widely cultivated and distributed.

Field Investigation and UAV Image Acquisition
Xingbin District is the only municipal district in Laibin City. A field survey was carried out in the south-central part of Xingbin District from December 7 to 10, 2017 (Figure 1). At that time, sugarcane entered the mature period and began to harvest gradually; most of the late rice has been harvested. From 10:00 to 13:00 on December 10, two consecutive hyperspectral images of UAV were acquired, covering an area of 230 × 330 m. A commercial snapshot hyperspectral imaging sensor (S185) onboard a multi-rotor UAV platform was used in this study. The S185 sensor employs two charge-coupled device (CCD) detectors with the 6.45 µm pixel size and the focal length of 23 mm. The S185 multi-rotor UAV system mainly includes: Cubert S185 hyperspectral data acquisition system, six-rotor electric UAV system, three-axis stabilized platform and data processing system. The S185 was radiometrically corrected with reference measurements on a white board and dark measurements by covering the black plastic lens cap prior to the flight. The reflectance was obtained by subtracting the dark measurement from the actual measured values during the flight and the reference values and then divided by this reference [33]. hyperspectral cubes with 12-bit radiometric resolution and with the size of 50 × 50 pixels were created. Accordingly, there were 898 panchromatic images with the size of 1000 × 1000 pixels. The typical vegetation types in the UAV coverage image are sugarcane, eucalyptus and citrus. In addition, there are also some vegetables, weeds and other vegetation types. All vegetation types are distributed alternately and extremely fragmented. A handheld Global Navigation Satellite System (GNSS) receiver (Juno SB, Trimble Navigation Limited, Sunnyvale, CA, USA) was used to record the position and category information of sample points, with a positioning accuracy of about 3 m. A total of 63 samples were recorded, which were evenly distributed throughout the study area.

Field Investigation and UAV Image Acquisition
Xingbin District is the only municipal district in Laibin City. A field survey was carried out in the south-central part of Xingbin District from December 7 to 10, 2017 (Figure 1). At that time, sugarcane entered the mature period and began to harvest gradually; most of the late rice has been harvested. From 10:00 to 13:00 on December 10, two consecutive hyperspectral images of UAV were acquired, covering an area of 230 × 330 m. A commercial snapshot hyperspectral imaging sensor (S185) onboard a multi-rotor UAV platform was used in this study. The S185 sensor employs two charge-coupled device (CCD) detectors with the 6.45 μm pixel size and the focal length of 23 mm. The S185 multi-rotor UAV system mainly includes: Cubert S185 hyperspectral data acquisition system, six-rotor electric UAV system, three-axis stabilized platform and data processing system. The S185 was radiometrically corrected with reference measurements on a white board and dark measurements by covering the black plastic lens cap prior to the flight. The reflectance was obtained by subtracting the dark measurement from the actual measured values during the flight and the reference values and then divided by this reference [33]. The measurement time to capture a full hyperspectral data cube is consistent with the automatic exposure time for white board reference measurement, about 1ms in a clear sky. With the flight height of 100 m, flight speed of 4.8 m/s, sampling interval of 0.8 s, heading overlap rate of about 80% and lateral overlap rate of about 70%, hyperspectral cubes of 125 effective bands (450-946 nm) and panchromatic images with the spatial resolution of 0.025 m were acquired synchronously. Totally 898 hyperspectral cubes with 12-bit

Preprocessing of UAV Hyperspectral Image
The preprocessing of S185 hyperspectral image mainly includes image fusion and mosaic [34]. Each hyperspectral cube and the corresponding panchromatic image synchronously collected were fused by using Cubert-Pilot software (Cubert GmbH, Ulm, Baden-Württemberg, Germany), and the fused hyperspectral cubes were obtained with a spatial resolution of 0.025 m [35]. The automatic image mosaic software Agisoft PhotoScan (Agisoft, St. Petersburg, Russia) was used for its stated accuracy 1-3 pixels and the high-quality mosaic [36]. By capturing overlapping images with a certain overlap rate, all the hyperspectral cubes were stitched together based on the point clouds of the panchromatic images [34]. Since the Position and Orientation System (POS) was not equipped onboard the UAV, the UAV hyperspectral image was geometrically corrected with reference to the 0.14 m Google Earth (GE) image dated 7 December 2016. As the spectrum less than 0.85 µm was consistent with that of vegetation, hyperspectral data in the range of 0.45-0.85 µm were used in the study, totaling 101 bands [37,38]. The corresponding wavelengths of 101 bands are available in Appendix A.
In order to explore the influence of spatial scale on vegetation classification of UAV images, the 0.025 m image was resampled to 0.05, 0.1, 0.25, 0.5, 1, 2.5 m, a total of 7 pixel scales according to the Remote Sens. 2020, 12, 146 5 of 25 spatial resolution range of most UAV images. As the smoothing or sharpening effects of the Nearest Neighbor, Bilinear and Cubic Convolution would affect the result analysis [39][40][41], the Pixel Aggregate was used to reduce the resolution of images. The Pixel Aggregate averages the resolution units into a larger one, avoiding the debate on the exact definition of spatial resolution and how to measure it [42], which has been widely used in multi-scale analysis [25,[43][44][45][46].

Reference Data
Based on the field observations and the recorded 63 field samples, a total of 59 polygons were manually outlined as training samples. The number of training polygons and corresponding pixels in multi-scale UAV images are shown in Table 1. In addition, a hand-drawn ground-truth reference image was delineated in detail for validation. Figure 2 illustrates the distribution of the reference image.

Methods
Based on UAV hyperspectral images of 7 different scales, sugarcane, eucalyptus, citrus and other vegetation were classified by using object-based image analysis (OBIA). Specifically, it included: (1) selecting appropriate scale for multi-resolution segmentation; (2) using the mean decrease accuracy (MDA) method for feature evaluation and selection; (3) selecting the classifiers of Support Vector Machines (SVM) and Random Forest (RF) to classify vegetation types of multi-scale images and comparing their accuracy differences; (4) analyzing the variation of appropriate segmentation parameters and feature space of multi-scale images, and discussing the influence of spatial scale variation on vegetation classification of UAV images.

Multi-Resolution Segmentation
Image segmentation is the first step of OBIA. It is a process of dividing an image into several discrete image objects (IOs) with unique properties according to certain criteria [47]. The accuracy of image segmentation significantly affects the accuracy of OBIA [48]. A bottom up region-merging technique based on the fractal net evolution algorithm proposed by Baatz and Schäpe was used for

Methods
Based on UAV hyperspectral images of 7 different scales, sugarcane, eucalyptus, citrus and other vegetation were classified by using object-based image analysis (OBIA). Specifically, it included: (1) selecting appropriate scale for multi-resolution segmentation; (2) using the mean decrease accuracy (MDA) method for feature evaluation and selection; (3) selecting the classifiers of Support Vector Machines (SVM) and Random Forest (RF) to classify vegetation types of multi-scale images and comparing their accuracy differences; (4) analyzing the variation of appropriate segmentation parameters and feature space of multi-scale images, and discussing the influence of spatial scale variation on vegetation classification of UAV images.

Multi-Resolution Segmentation
Image segmentation is the first step of OBIA. It is a process of dividing an image into several discrete image objects (IOs) with unique properties according to certain criteria [47]. The accuracy of image segmentation significantly affects the accuracy of OBIA [48]. A bottom up region-merging technique based on the fractal net evolution algorithm proposed by Baatz and Schäpe was used for multi-resolution segmentation [49]. As the most widely used method, it can generate highly homogeneous segmentation regions, thus separating and representing ground objects in the best scale [50]. In the study, two segmentations were performed, creating two levels of IOs. A series of interactive "trial and error" tests were used to determine the proper segmentation parameters [51,52]. Six spectral bands including blue, green, red, red edge I, red edge II and near-infrared were used as inputs. A small-scale factor was used for the first segmentation to maximize the separation of vegetation and non-vegetation. In the second segmentation, three scale factors were set separately to prevent the objects being too fragmented and the appropriate segmentation of the three was determined by the classification accuracy.

Feature Extraction
In the process of OBIA, the features related to the IOs can be extracted from the UAV image. The ideal features should reflect the differences between the target types.
Sugarcane, citrus and eucalyptus are broadleaf vegetation, showing different spatial distribution features in high spatial resolution images. Sugarcane stalks are 3-5 m high, with clumps of leaves. The leaves are about 1m long and 4-6 cm wide, the edges of which are serrated and rough. The rows are 80-100 cm apart, showing the characteristics of fine and compact distribution. Citrus is an evergreen small tree with a height of about 2 m and a round crown of less than 2 m. The plant spacing is 1.5-2 m. The leaves are ovate-lanceolate, with a large size variation and a length of 4-8 cm. It shows a feature of regular, sparse and circular crown distribution. Eucalyptus is an evergreen dense-shade tall tree with a height of 20 m. Its crown is triangular spire-shaped, with small crown and opposite leaves in heart or broadly lanceolate shape, showing a feature of dense distribution.

Feature Evaluation and Reduction
High-dimensional data usually need feature selection before machine learning. The significance of feature selection lies in reducing data redundancy, strengthening the understanding of features, enhancing the generalization ability of models and improving the processing efficiency. Random Forest (RF) can effectively reduce the data dimension while ensuring the classification accuracy, which is a machine learning algorithm composed of multiple classification and regression trees (CART) proposed by Breiman [71]. It is widely used in classification and identification of images and selection of high-dimensional features [72,73]. The mean decrease accuracy (MDA) method of RF was adopted for feature importance evaluation, which could disturb the eigenvalue order of each feature and then evaluate the importance of the feature by measuring the influence of this change on the accuracy of the model. If a feature is important, its order change will significantly reduce the accuracy.
On the basis of MDA results, all the features are ranked from big to small according to the feature importance, and different number of features will be used successively to classify vegetation types. In order to eliminate feature redundancy that might be caused by repeated participation of adjacent 101 reflectance bands, the following feature reduction principle is adopted based on feature importance evaluation: For 101 bands, they are first ranked in order of feature importance from big to small. When a band is retained, the two adjacent bands above and below will be deleted.
Sometimes, it may occur that the interval between the band to be retained and the band already retained is 3, which is also acceptable. However, if the interval is 4, this band should be deleted, and its upper or lower band should be retained to ensure that the retained band has relatively strong importance and the number of bands deleted between the retained bands is 2 or 3.
For example, if the first important band is the 64th band, it should be retained first, and the 62nd, 63rd, 65th and 66th bands should be deleted at the same time. If the second important band is the 61st or 60th band, both cases are acceptable because it can ensure that the number of deleted bands is 2 or 3. However, if the second important band is the 59th band, it needs to be deleted, and the more important one of the 58th or 60th band needs to be retained according to the importance of the two.

Classifier
Two supervised classifiers were considered in this paper: Support Vector Machines (SVM) and Random Forest (RF).
SVM is a non-parametric classifier [74]. Adopting the principle of structural risk minimization, it automatically selects important data points to support decision-making. It provides a brand-new solution for classification of ground objects in high-resolution images with extraordinary efficiency. In this study, we used Radial Basis Function Kernel (RBF) function because of its outperformance in classification [75]. Two important classifier parameters need to be determined, including RBF kernel parameter gamma and penalty factor c. A moderate error penalty value (c = 100) and a gamma equaling to the inverse of the feature dimension were configured to facilitate the comparison between the different classification results related to different feature numbers [76,77].
As mentioned above, RF is a non-parametric ensemble learning algorithm [71], which is composed of multiple decision-making trees. In the process of building decision-making trees, the splitting of each point is determined by Gini coefficient criterion to realize the best variable splitting. RF has the characteristics of fast learning speed, strong robustness and generalization ability. It has the abilities to analyze and select complex interaction features, being widely used in computer vision, human identification, image processing and other fields. Two important parameters need to be determined, including ntree (the number of decision trees executing classification) and mtry (the number of input variables used at each node) [78]. Different numbers of ntree were tested from 50 to 150 at 10 tree intervals. Classification accuracy did not change much as the number changed. Consistently used ntree (ntree = 50) and mtry equaling to the square root of the feature dimension were configured to facilitate the comparison between the different classification results related to different feature numbers.

Results of Multi-Resolution Segmentation and Vegetation Information Extraction
Two segmentations were performed by multi-resolution segmentation. As the most important segmentation parameter, segmentation scale greatly affects the identification results of vegetation types. In order to distinguish vegetation information from non-vegetation to the greatest extent, the scale of the first segmentation should be as small as possible so that the segmentation object would not contain both vegetation and non-vegetation information (Table 4). For multi-scale images, six bands including blue (band1), green (band26), red (band56), red edge I (band65), red edge II (band74) and near-infrared (band88) were used as inputs, and the weight of each band was 1. By the "trial and error" method and previous study experience [79,80], the first segmentation was completed under the spectral parameter of 0.9, shape parameter of 0.1, compactness of 0.5 and smoothness of 0.5. NDVI and mNDVI705 were jointly applied to extract vegetation information. NDVI can reflect the distribution density and grow conditions of vegetation to the maximum extent, so it is often regarded as the most important vegetation index for vegetation information extraction. However, when NDVI was used alone to extract vegetation information, a few marginal areas of vegetation, ridges of field and harvested rice stubbles were confused with vegetation information, so mNDVI705 was added to help the extraction of vegetation information. Derived from NDVI705, mNDVI705 takes the specular reflection characteristics of leaves into consideration and is good at capturing the subtle characteristics of leaf senescence and small changes in leaf canopy. Thus, it has been widely used in fine agriculture, forest monitoring and vegetation stress monitoring [81][82][83]. The multi-parameter threshold method was adopted, i.e., NDVI and mNDVI705 respectively take thresholds. When NDVI was greater than or equal to 0.29 and mNDVI705 was greater than or equal to 0.256, vegetation information in the study area would be accurately extracted.
On the basis of vegetation information that had been extracted, the objects of vegetation information were segmented for the second time, in order to utilize the textural features of objects and ensure that the classification results are not excessively fragmented. And then, citrus, sugarcane, eucalyptus and other vegetation were identified on this basis. Three different segmentation scales were respectively set for multi-scale images, and then the appropriate segmentation scale for multi-scale images was determined by the classification accuracy (Table 5). Taking the secondary segmentation result of 0.25 m image in Figure 3d,e,f at three scales as an example, when the parameter of segmentation scale is too small, the objects cannot reflect rich textural information. If it is appropriate, the textural structure can be effectively reflected without being too fragmented. Once it is too big, for vegetation types with similar spatial and spectral features, the probability of objects containing different vegetation types will increase. characteristics of leaf senescence and small changes in leaf canopy. Thus, it has been widely used in fine agriculture, forest monitoring and vegetation stress monitoring [81][82][83]. The multi-parameter threshold method was adopted, i.e., NDVI and mNDVI705 respectively take thresholds. When NDVI was greater than or equal to 0.29 and mNDVI705 was greater than or equal to 0.256, vegetation information in the study area would be accurately extracted.
On the basis of vegetation information that had been extracted, the objects of vegetation information were segmented for the second time, in order to utilize the textural features of objects and ensure that the classification results are not excessively fragmented. And then, citrus, sugarcane, eucalyptus and other vegetation were identified on this basis. Three different segmentation scales were respectively set for multi-scale images, and then the appropriate segmentation scale for multiscale images was determined by the classification accuracy (Table 5). Taking the secondary segmentation result of 0.25 m image in Figure 3d,e,f at three scales as an example, when the parameter of segmentation scale is too small, the objects cannot reflect rich textural information. If it is appropriate, the textural structure can be effectively reflected without being too fragmented. Once it is too big, for vegetation types with similar spatial and spectral features, the probability of objects containing different vegetation types will increase.
With the decrease of spatial resolution, the occurrence probability of mixed pixels increases. In order to ensure the homogeneity of vegetation types in the same object, the segmentation scale gradually decreases. However, the appropriate segmentation scale does not show a simple recursive variation with spatial resolution or data volume. For images with 0.1-2.5 m spatial resolution, the segmentation scale is concentrated in 25-100, while the segmentation scales for 0.05 and 0.025 m images rapidly rise to 1000 and 2500, which shows that the ability of the classifier to correctly recognize larger objects is enhanced in images with centimeter-level resolution. (g) (h) (i)

Importance Variation of Different Feature Types
To evaluate the feature importance variation of multi-scale images and rank the features involved in the classification, the importance of 149 features was measured by MDA method. In the process of MDA, based on the 149 features calculated for the training IOs derived from the training rectangles, the importance scores of 7 multi-scale images in each second segmentation scale were obtained, respectively. The feature importance was separately analyzed in the categories of spectrum, vegetation index, texture and shape for multi-scale images under the condition of the appropriate segmentation scale.
1. Importance Analysis of Spectral Features in Multi-Scale Images According to the normalized measurement results of importance for 101 reflectance features in multi-scale images and their boxplot distribution, as shown in Figure 4a, for 0.025-0.5 m images, the important spectral ranges are located in red edge I, followed by red, green and blue regions. With the decrease of spatial resolution, the importance of red edge I for 1-2.5 m images decreases, and the green and yellow bands increase in importance.
According to the normalized measurement results of importance for other spectral features in multi-scale images, as shown in Figure 4b, Ratio65 is in the forefront of importance evaluation in all scale images, i.e., the contribution ratio of red edge I is in the important position. For 0.025-0.5 m images, the features of StdDev are of great importance, but their importance declines rapidly with the decrease of spatial resolution. The importance of intensity, hue and brightness of 2.5 m image increases.
2. Importance Analysis of Vegetation Index Features in Multi-Scale Images As shown in Figure 4c, with the decrease of spatial resolution, the importance of vegetation index features increases. The vegetation indices in the forefront of importance evaluation are all related to the red edge. Among broadband indices that can be calculated based on traditional 4-band multi-spectral images, VIgreen is of greater importance in all scale images; and NDVI, the most widely used index, is of greater importance in 0.1 m image.
3. Importance Analysis of Shape and Textural Features in Multi-Scale Images As shown in Figure 4d, with the decrease of spatial resolution, the importance of textural features increases slightly in fluctuation and then decreases. For 0.025-0.5 m images, GLCM ASM, GLCM Entropy and GLCM Mean are stable and in the forefront of importance evaluation. For 0.1-0.5 m images, GLDV ASM and GLDV Entropy are more important. In general, the importance of textural features for 1 and 2.5 m images is weakened, which is reflected in the reduction of pixel numbers contained in low-resolution images for the object scale, resulting in textural difference weakening of different vegetation types in low-resolution images. Shape features are not of sufficient importance in multi-scale images, and only the 2.5 m image is of strong importance for the compactness feature. With the decrease of spatial resolution, the occurrence probability of mixed pixels increases. In order to ensure the homogeneity of vegetation types in the same object, the segmentation scale gradually decreases. However, the appropriate segmentation scale does not show a simple recursive variation with spatial resolution or data volume. For images with 0.1-2.5 m spatial resolution, the segmentation scale is concentrated in 25-100, while the segmentation scales for 0.05 and 0.025 m images rapidly rise to 1000 and 2500, which shows that the ability of the classifier to correctly recognize larger objects is enhanced in images with centimeter-level resolution.

Importance Variation of Different Feature Types
To evaluate the feature importance variation of multi-scale images and rank the features involved in the classification, the importance of 149 features was measured by MDA method. In the process of MDA, based on the 149 features calculated for the training IOs derived from the training rectangles, the importance scores of 7 multi-scale images in each second segmentation scale were obtained, respectively. The feature importance was separately analyzed in the categories of spectrum, vegetation index, texture and shape for multi-scale images under the condition of the appropriate segmentation scale.

Importance Analysis of Spectral Features in Multi-Scale Images
According to the normalized measurement results of importance for 101 reflectance features in multi-scale images and their boxplot distribution, as shown in Figure 4a, for 0.025-0.5 m images, the important spectral ranges are located in red edge I, followed by red, green and blue regions. With the decrease of spatial resolution, the importance of red edge I for 1-2.5 m images decreases, and the green and yellow bands increase in importance.
According to the normalized measurement results of importance for other spectral features in multi-scale images, as shown in Figure 4b, Ratio65 is in the forefront of importance evaluation in all scale images, i.e., the contribution ratio of red edge I is in the important position. For 0.025-0.5 m images, the features of StdDev are of great importance, but their importance declines rapidly with the decrease of spatial resolution. The importance of intensity, hue and brightness of 2.5 m image increases.

Importance Analysis of Vegetation Index Features in Multi-Scale Images
As shown in Figure 4c, with the decrease of spatial resolution, the importance of vegetation index features increases. The vegetation indices in the forefront of importance evaluation are all related to the red edge. Among broadband indices that can be calculated based on traditional 4-band multi-spectral images, VIgreen is of greater importance in all scale images; and NDVI, the most widely used index, is of greater importance in 0.1 m image.

Importance Analysis of Shape and Textural Features in Multi-Scale Images
As shown in Figure 4d, with the decrease of spatial resolution, the importance of textural features increases slightly in fluctuation and then decreases. For 0.025-0.5 m images, GLCM ASM, GLCM Entropy and GLCM Mean are stable and in the forefront of importance evaluation. For 0.1-0.5 m images, GLDV ASM and GLDV Entropy are more important. In general, the importance of textural features for 1 and 2.5 m images is weakened, which is reflected in the reduction of pixel numbers contained in low-resolution images for the object scale, resulting in textural difference weakening of different vegetation types in low-resolution images. Shape features are not of sufficient importance in multi-scale images, and only the 2.5 m image is of strong importance for the compactness feature.

Variation of Feature Types within Different Feature Groups
In order to eliminate feature redundancy that might be caused by repeated participation in classification of adjacent 101 reflectance bands, the feature reduction principle in Section 3.3 was adopted. Due to the different feature importance order of 101 bands for the 7 scale images, accordingly, the number of deleted bands among the retained bands was 2 or 3, which was not exactly the same, resulting in a slight difference in the number of retained bands. After the feature reduction of 101 bands, the number of retained bands was 29-31 for multi-scale images. Together with other spectral, vegetation index, textural and shape features, the total number of features after reduction was 77-79. In the following analysis, the value of "80 features" was used instead of "77-79" for uniform marking.
Finally, according to the importance from big to small, the final features were divided into 8 groups with 10 features in each group. The number of features for different feature types in each group was divided by the total number of this feature type. The results show that the proportion of different feature types in each group is different ( Figure 5). In general, spectral features are distributed in all groups with little difference. The proportion of vegetation index features for multi-

Variation of Feature Types within Different Feature Groups
In order to eliminate feature redundancy that might be caused by repeated participation in classification of adjacent 101 reflectance bands, the feature reduction principle in Section 3.3 was adopted. Due to the different feature importance order of 101 bands for the 7 scale images, accordingly, the number of deleted bands among the retained bands was 2 or 3, which was not exactly the same, resulting in a slight difference in the number of retained bands. After the feature reduction of 101 bands, the number of retained bands was 29-31 for multi-scale images. Together with other spectral, vegetation index, textural and shape features, the total number of features after reduction was 77-79. In the following analysis, the value of "80 features" was used instead of "77-79" for uniform marking.
Finally, according to the importance from big to small, the final features were divided into 8 groups with 10 features in each group. The number of features for different feature types in each group was divided by the total number of this feature type. The results show that the proportion of different feature types in each group is different ( Figure 5). In general, spectral features are distributed in all groups with little difference. The proportion of vegetation index features for multi-scale images in each group is quite different, which shows that the importance of vegetation index features increases as the spatial resolution decreases. Textural features are mainly distributed in the last 4 groups. In the first 4 groups, textural features are mainly distributed in images with high spatial resolution, while the textures in low-resolution images are not of sufficient importance. Shape features perform poorly in each scale image and are mainly distributed in the latter 3 groups. Compared to images with high spatial resolution, shape features in low-resolution images are enhanced. These results are consistent with previous studies [8,54,78,84]. In these studies, the spectral and vegetation index features make more significant contributions to the species classification of high spatial resolution images than the textural and shape features. Specifically, only a small percentage of texture or shape features are selected for classification. The majorities and most important features are spectral and vegetation index features.
Remote Sens. 2020, 12, 146 13 of 26 features perform poorly in each scale image and are mainly distributed in the latter 3 groups. Compared to images with high spatial resolution, shape features in low-resolution images are enhanced. These results are consistent with previous studies [8,54,78,84]. In these studies, the spectral and vegetation index features make more significant contributions to the species classification of high spatial resolution images than the textural and shape features. Specifically, only a small percentage of texture or shape features are selected for classification. The majorities and most important features are spectral and vegetation index features.

Classification Results
According to the results of feature importance evaluation and selection, SVM classifier was used to classify multi-scale images by using different segmentation scales and different number of features successively, adding 10 features each time. This means that the first 10 features will be used for the first classification and 80 features will be used for the eighth classification. The overall accuracy (OA) and Kappa coefficient were used to evaluate the classification accuracy based on the ground-truth reference image covering the whole study area [85]. Figure 6 presents the classification results of multi-scale images based on SVM classifier under the condition of appropriate segmentation scale and feature number.
RF classifier was used to perform classification to verify the results of SVM classifier. As shown in Table 6, compared with the classification results of vegetation type for multi-scale images based on SVM classifier under appropriate segmentation scale and feature number, the results of RF classifier under corresponding conditions were slightly lower than those of SVM, and the OA decreased by 1.3%-4.2%, which was consistent with previous study results [86][87][88]. In these studies, SVM classifier has higher classification accuracy and stronger robustness than RF. Furthermore, classification experiments based on all 149 features were carried out by SVM and RF classifiers. Compared to the results using all 149 features, it shows that the classification accuracy of SVM classifier using the appropriate number of features has been greatly improved, with the OA increasing by 2.5%-8.3%. These results are consistent with previous studies; that is, for too many spectral bands or target features, the classification performance and efficiency can be improved by eliminating redundant information [89,90]. It is worth noting that RF is insensitive to the number of features and could provide better results than SVM when using all 149 features [90].
In addition, the pixel-based classification was performed based on 101 reflectance bands using

Classification Results
According to the results of feature importance evaluation and selection, SVM classifier was used to classify multi-scale images by using different segmentation scales and different number of features successively, adding 10 features each time. This means that the first 10 features will be used for the first classification and 80 features will be used for the eighth classification. The overall accuracy (OA) and Kappa coefficient were used to evaluate the classification accuracy based on the ground-truth reference image covering the whole study area [85]. Figure 6 presents the classification results of multi-scale images based on SVM classifier under the condition of appropriate segmentation scale and feature number.
RF classifier was used to perform classification to verify the results of SVM classifier. As shown in Table 6, compared with the classification results of vegetation type for multi-scale images based on SVM classifier under appropriate segmentation scale and feature number, the results of RF classifier under corresponding conditions were slightly lower than those of SVM, and the OA decreased by 1.3%-4.2%, which was consistent with previous study results [86][87][88]. In these studies, SVM classifier has higher classification accuracy and stronger robustness than RF. Furthermore, classification experiments based on all 149 features were carried out by SVM and RF classifiers. Compared to the results using all 149 features, it shows that the classification accuracy of SVM classifier using the appropriate number of features has been greatly improved, with the OA increasing by 2.5%-8.3%. These results are consistent with previous studies; that is, for too many spectral bands or target features, the classification performance and efficiency can be improved by eliminating redundant information [89,90]. It is worth noting that RF is insensitive to the number of features and could provide better results than SVM when using all 149 features [90]. In addition, the pixel-based classification was performed based on 101 reflectance bands using SVM classifier (Table 6)   (g) Figure 6. Classification results of the study area for multi-scale UAV images based on SVM classifier. The UAV image and the ground truth reference image of the study area are available in Figure 1 and  Figure 7 shows the OA variation for multi-scale images using SVM classifier in the condition of different segmentation scales and feature numbers. It can be seen that segmentation scales and feature numbers have great influence on the classification accuracy of multi-scale images, further confirming the importance of segmentation scale optimization and multi-feature analysis.  Figure 7 shows the OA variation for multi-scale images using SVM classifier in the condition of different segmentation scales and feature numbers. It can be seen that segmentation scales and feature numbers have great influence on the classification accuracy of multi-scale images, further confirming the importance of segmentation scale optimization and multi-feature analysis.

Overall Accuracy Variation of Multi-Scale Images
Remote Sens. 2020, 12, 146 16 of 26 (g) Figure 6. Classification results of the study area for multi-scale UAV images based on SVM classifier. The UAV image and the ground truth reference image of the study area are available in Figure 1 and  Figure 7 shows the OA variation for multi-scale images using SVM classifier in the condition of different segmentation scales and feature numbers. It can be seen that segmentation scales and feature numbers have great influence on the classification accuracy of multi-scale images, further confirming the importance of segmentation scale optimization and multi-feature analysis. As shown in Figure 8, the number of appropriate features for multi-scale images is 30-40. With the increase of feature number, the identification accuracy shows a trend of increasing first and then decreasing for multi-scale images, which reveals that too few features cannot achieve the purpose of high-precision identification of vegetation types, while too many features have strong correlation and information redundancy, resulting in interference information for vegetation classification. As shown in Figure 8, the number of appropriate features for multi-scale images is 30-40. With the increase of feature number, the identification accuracy shows a trend of increasing first and then decreasing for multi-scale images, which reveals that too few features cannot achieve the purpose of high-precision identification of vegetation types, while too many features have strong correlation and information redundancy, resulting in interference information for vegetation classification.  Figure 9 shows the OA variation under the condition of appropriate feature number and segmentation scale with different spatial scale by using SVM classifier. As the spatial resolution decreases, the OA shows a stable and slight fluctuation, and then gradually decreases. Among them, the accuracies of 0.025, 0.05, 0.1 and 0.25 m images showed little difference and reached more than 90%. The accuracy of 0.05 m image was the highest, reaching 91.3%, followed by 0.25 m image with an accuracy of 91.0%. Since the spatial resolution of 0.5 m, the OA gradually declined from 88.6% to 84.3% with the decrease of spatial resolution. The classification accuracy of images at various scales is consistent with previous studies. For example, Underwood used 4 m AVIRIS hyperspectral image, and the classification accuracy for six communities of three invasive species reaches 75% [28]. In Cao's study, the classification accuracy for mangrove species based on 0.15, 0.3 and 0.5 m UAV hyperspectral images reach 88.66%, 86.57% and 82.69%, respectively [91].

Overall Accuracy Variation of Multi-Scale Images
In this case study, the OA drops if the resolution is coarser than 0.25 m. With the decrease of spatial resolution, the number of mixed pixels increase continuously, and the edge of the parcels is more likely to increase commission and omission errors. This means that the spatial resolution should reach a certain threshold to achieve decent accuracy. However, same as previous cognition, the finer the spatial resolution is not the better [31,92]. For example, the accuracy of 0.025 m image is slightly lower than that of 0.05 m and 0.25 m. This makes sense, as vegetation has specific physical size, and spatial resolution significantly finer than the certain threshold may not help the classification performance. The image with centimeter-level resolution helps us not only to understand the detailed information of vegetation types but also significantly strengthens the phenomenon of same objects with different spectrum, bringing more interference to vegetation type identification. At the same time, the ultrahigh resolution has resulted in multiple increases in information, significantly reduces the efficiency of image processing. It means that there is no need to excessively pursue finer resolution than the threshold. A low-cost imager may be sufficient. Alternatively, the UAV can fly at higher altitude with larger coverage by sacrificing the spatial resolution, which is still sufficient for vegetation classification.  Figure 9 shows the OA variation under the condition of appropriate feature number and segmentation scale with different spatial scale by using SVM classifier. As the spatial resolution decreases, the OA shows a stable and slight fluctuation, and then gradually decreases. Among them, the accuracies of 0.025, 0.05, 0.1 and 0.25 m images showed little difference and reached more than 90%. The accuracy of 0.05 m image was the highest, reaching 91.3%, followed by 0.25 m image with an accuracy of 91.0%. Since the spatial resolution of 0.5 m, the OA gradually declined from 88.6% to 84.3% with the decrease of spatial resolution. The classification accuracy of images at various scales is consistent with previous studies. For example, Underwood used 4 m AVIRIS hyperspectral image, and the classification accuracy for six communities of three invasive species reaches 75% [28]. In Cao's study, the classification accuracy for mangrove species based on 0.15, 0.3 and 0.5 m UAV hyperspectral images reach 88.66%, 86.57% and 82.69%, respectively [91].

Identification Accuracy Variation of Each Vegetation Type
Producer's accuracy (PA), user's accuracy (UA) and F-Score were used to evaluate the classification accuracy of each vegetation type, and the maximum F-score of the results based on different feature numbers was taken as the best identification accuracy of each vegetation type at various scales.
As shown in Figures 10 and 11, the maximum F-score of sugarcane is the highest, reaching In this case study, the OA drops if the resolution is coarser than 0.25 m. With the decrease of spatial resolution, the number of mixed pixels increase continuously, and the edge of the parcels is more likely to increase commission and omission errors. This means that the spatial resolution should reach a certain threshold to achieve decent accuracy. However, same as previous cognition, the finer the spatial resolution is not the better [31,92]. For example, the accuracy of 0.025 m image is slightly lower than that of 0.05 m and 0.25 m. This makes sense, as vegetation has specific physical size, and spatial resolution significantly finer than the certain threshold may not help the classification performance. The image with centimeter-level resolution helps us not only to understand the detailed information of vegetation types but also significantly strengthens the phenomenon of same objects with different spectrum, bringing more interference to vegetation type identification. At the same time, the ultrahigh resolution has resulted in multiple increases in information, significantly reduces the efficiency of image processing. It means that there is no need to excessively pursue finer resolution than the threshold. A low-cost imager may be sufficient. Alternatively, the UAV can fly at higher altitude with larger coverage by sacrificing the spatial resolution, which is still sufficient for vegetation classification.

Identification Accuracy Variation of Each Vegetation Type
Producer's accuracy (PA), user's accuracy (UA) and F-Score were used to evaluate the classification accuracy of each vegetation type, and the maximum F-score of the results based on different feature numbers was taken as the best identification accuracy of each vegetation type at various scales.
As shown in Figures 10 and 11, the maximum F-score of sugarcane is the highest, reaching 92.4%-95.5% for multi-scale images. The 0.025-1 m images are not sensitive to the variation of feature numbers. The PA of sugarcane varies greatly with different feature numbers of the 2.5 m image, i.e., with the decrease of spatial resolution, the importance of feature selection increases. The maximum F-score of eucalyptus is between 82.0% and 90.4% in various scales. Except for the 2.5 m image, the variation of feature number has great influence on the identification of eucalyptus. The maximum F-score of citrus in each scale image is poorly. When 40 features of 0.25 m image participate in classification, the F-score reached the maximum, 63.39%. Then with the decrease of spatial resolution, the identification accuracy decreases greatly. The PA of citrus is relatively acceptable, while the UA performs terribly, which shows that as much citrus information as possible is extracted, while more non-citrus is identified as citrus. In addition, the variation of feature numbers has great influence on the accurate identification of citrus. Similar to citrus, the type of other vegetation receives low accuracy. When 30 features of 0.025 m image participated in classification, the F-score reached the maximum, 66.69%. Since the 0.25 m image, the identification accuracy gradually declines with the decrease of spatial resolution.
In general, although the parcels of sugarcane and eucalyptus are irregular in shape in multi-scale images, most parcels have an area of more than 125 m 2 , accounting for about 20 pixels in 2.5 m image. Moreover, they are significantly different from citrus and other vegetation and are easier to be identified with accurate and stable accuracy for all 7 scales. However, citrus and other vegetation acquired unsatisfactory results, and the identification accuracies decreased greatly since the resolution of 0.25 m. Citrus plants are small, and the planting structure is sparse. The crown boundary can be segmented in images with spatial resolution better than 1m. With the decrease of spatial resolution, dwarf citrus plants and surrounding ground objects will be formed into mixed pixels. It is difficult to segment the crown boundary based on 2.5 m image. In addition, other vegetation scattered in the study area tend to be confused with citrus, resulting in the reduction of the identification accuracies for citrus and other vegetation. At the same time, the edge of each vegetation type is mostly mixed or scattered with the type of other vegetation, and the commission and omission in the classification results also occur at the edge of different vegetation types, which shows that the occurrence of mixed pixels in the transition region of different vegetation types increases with the decrease of spatial resolution, bringing greater uncertainty to the classification results of vegetation types.

Conclusions
This study aims to evaluate the impact of spatial resolution on the classification of vegetation types in highly fragmented planting areas based on UAV hyperspectral images. By aggregating the 0.025 m UAV hyperspectral image into coarse spatial resolutions (0.05, 0.1, 0.25, 0.5, 1, 2.5 m), we have simulated centimeter-to-meter level resolution images that can be obtained by the UAV system, and evaluated the accuracy variation of the fine classification of several vegetation types such as sugarcane, citrus and eucalyptus in southern China based on multi-scale images. The results show that the classification accuracy of vegetation types is closely related to the scale of remote sensing images. For this study area, with the decrease of spatial resolution, the OA shows a stable and slight fluctuation and then gradually decreases. The best classification accuracy does not occur in the original image but at an intermediate level of resolution. These results are consistent with similar studies on image scale, i.e., the best resolution occurs when the spectral intra-class variance is the smallest, and the class has not yet begun to mix spatially. Therefore, the ideal spatial resolution should vary according to the diversity and distribution of species in the ecosystem. Parcel size and distribution are the key factors that determine the accuracy at a given resolution. Due to the existence of small and fragmented parcels, images with coarse resolution no longer contain some original categories, such as citrus in this study, resulting in the reduction of classification accuracy of 1 and 2.5 m images. Therefore, it is important to select images of appropriate spatial scale according to the special distribution and parcel size of the study area so as to obtain more ideal classification accuracy in UAV flight experiment, data processing and application. In the process of OBIA, based on the results of multi-feature evaluation and analysis, it is successful to classify vegetation types of images in different scales by using different feature numbers and segmentation parameters. We find that with the decrease of the spatial resolution, the importance of vegetation index features increases and that of textural features shows an opposite trend; the appropriate segmentation scale decreases gradually, and the appropriate number of features is 30-40, which means that the feature parameters vary for multi-scale images. Therefore, appropriate feature parameters need to be selected for images in different scales to ensure the accuracy of classification.

Conclusions
This study aims to evaluate the impact of spatial resolution on the classification of vegetation types in highly fragmented planting areas based on UAV hyperspectral images. By aggregating the 0.025 m UAV hyperspectral image into coarse spatial resolutions (0.05, 0.1, 0.25, 0.5, 1, 2.5 m), we have simulated centimeter-to-meter level resolution images that can be obtained by the UAV system, and evaluated the accuracy variation of the fine classification of several vegetation types such as sugarcane, citrus and eucalyptus in southern China based on multi-scale images. The results show that the classification accuracy of vegetation types is closely related to the scale of remote sensing images. For this study area, with the decrease of spatial resolution, the OA shows a stable and slight fluctuation and then gradually decreases. The best classification accuracy does not occur in the original image but at an intermediate level of resolution. These results are consistent with similar studies on image scale, i.e., the best resolution occurs when the spectral intra-class variance is the smallest, and the class has not yet begun to mix spatially. Therefore, the ideal spatial resolution should vary according to the diversity and distribution of species in the ecosystem. Parcel size and distribution are the key factors that determine the accuracy at a given resolution. Due to the existence of small and fragmented parcels, images with coarse resolution no longer contain some original categories, such as citrus in this study, resulting in the reduction of classification accuracy of 1 and 2.5 m images. Therefore, it is important to select images of appropriate spatial scale according to the special distribution and parcel size of the study area so as to obtain more ideal classification accuracy in UAV flight experiment, data processing and application. In the process of OBIA, based on the results of multi-feature evaluation and analysis, it is successful to classify vegetation types of images in different scales by using different feature numbers and segmentation parameters. We find that with the decrease of the spatial resolution, the importance of vegetation index features increases and that of textural features shows an opposite trend; the appropriate segmentation scale decreases gradually, and the appropriate number of features is 30-40, which means that the feature parameters vary for multi-scale images. Therefore, appropriate feature parameters need to be selected for images in different scales to ensure the accuracy of classification.
There are several clear directions for future study. First, a more realistic simulation and amplification for images of fine spatial resolutions will help to improve the evaluation of potential applications of similar data in coarse resolution. The study on up-scaling of remote sensing images also shows that the spectral information of the resampled images has a strong dependence on the original images, resulting in differences with the actual observation results at a given scale [93]. The results of this study should be compared with the classification results using actual observation images for further understanding the potential impact of resolution on the classification of vegetation types. In addition, the improvement of spatial resolution will lead to greater intra-class difference and inter-class similarity, which will usually result in classification errors [94,95]. In view of the challenges and potential of ultrahigh resolution UAV images in the classification of vegetation types, advanced data analysis technologies developed in computer vision and machine learning, such as deep learning [96], should be comprehensively analyzed to improve the application capability of UAV images.
Author Contributions: Conceptualization, J.Y., T.Y. and X.G.; methodology, M.L.; investigation and data acquisition, M.L. and Z.S.; data analysis and original draft preparation, M.L., Z.Z. and X.M.; validation and writing-review and editing, Z.S., W.C. and J.L.; All authors contributed to the discussion, provided suggestions to improve the manuscript and checked the writing. All authors have read and agreed to the published version of the manuscript.
Funding: This study was funded by the Guangxi Science and Technology Development Project of Major Projects (Guike AA18118048-2) and the National Civil Space Infrastructure Project (17QFGW02KJ).

Conflicts of Interest:
The authors declare no conflict of interest.