The Sensitivity Feature Analysis for Tree Species Based on Image Statistical Properties

: While the statistical properties of images are vital in forestry engineering, the usefulness of these properties in various forestry tasks may vary, and certain image properties might not be enough to adequately describe a particular tree species. To address this problem, we propose a novel method to comprehensively analyze the relationship between various image statistical properties and images of different tree species, and to determine the subset of features that best describe each individual tree species. In this study, we employed various image statistical properties to quantify images of ﬁve distinct tree species from diverse places. Multiple feature-ﬁltering methods were used to ﬁnd the feature subset with the greatest correlation with the tree species category variable. Support Vector Machines (SVM) were employed to determine the number of features with the greatest correlation with the tree species, and a grid search was used to optimize the model. For each type of tree species image, we obtained the important ranking of all features in this type of tree species, and the sensitive feature subset of various tree species according to the order of features was determined by adding them to the Deep Support Vector Data Description (Deep SVDD). Finally, the feasibility of using a sensitive subset of the tree species was conﬁrmed. The experimental results revealed that by utilizing the ﬁltering method in conjunction with SVM, a total of eight feature subsets with the highest correlation with tree species categories were identiﬁed. Additionally, the sensitive feature subsets of different tree species exhibited signiﬁcant differences. Remarkably, employing the sensitive feature subset of each tree species resulted in F1-score higher than 0.7 for all tree species. These experimental results demonstrate that the sensitive feature subset of tree species based on image statistical properties can serve as a potential representation of a speciﬁc tree species, while features that are less strongly associated with tree species may be signiﬁcant in related areas, such as forestry protection and other related ﬁelds.


Introduction
Forests are crucial components of our ecosystem and contribute significantly to upholding ecological equilibrium [1].Studies indicate that employing the statistical properties of images can differentiate forest landscapes and enhance the accuracy of forest reconstruction [2].These findings laid the foundation for efficient forest resource data collection and intelligent operations of forestry robots [3].Furthermore, statistical properties demonstrating significant relationships can be utilized for purposes such as detecting forest fires [4], improving forest tree breeding [5], and other related applications in the field of forestry engineering.
Trees play a crucial role in maintaining ecological balance as they are an integral component of forest ecosystems [6].By providing a quantitative representation of tree species images through the application of statistical properties, image data restrictions due to subjective visual interpretation can be overcome [7], which helps to improve our Forests 2023, 14, 1057 2 of 17 understanding of forests.Notably, the specific statistical properties of images exhibit strong correlations with various tree species [8,9].For instance, the texture property of an image can reveal the smoothness or roughness of a tree's bark, which significantly affects its resistance to environmental stressors, such as pests and diseases [10].Wang et al. combined texture and differential spectral features to enhance the accuracy of mangrove species classification, which greatly improved the mapping accuracy [11].Assessing the sensitivity features of different tree species can be challenging due to the highly similar image statistical properties shared by different tree species, which can be attributed to their common ecological features [12].However, not all image statistical properties are equally important in different tree species, and some statistical properties that are less connected with tree species are more applicable to other forestry activities.
In recent years, with the rapid advancement of computer vision and data analysis, the field of utilizing machine learning to analyze tree species has gained momentum [13,14].The one-class sensitive feature subset refers to the subset of features that can distinguish this category from other categories for a specific category in the data set.Tree species sensitivity features represent the properties that are most indicative of the uniqueness of a particular tree species.At present, the use of one-class classification to find a subset of sensitive features of a specific category is mainly used in the diagnosis of different crop diseases [15], identification of invasive tree species [16], and rolling bearing fault diagnosis [17].However, adopting a kernel-based sensitive feature subset selected for a specific class of study may still provide challenges due to the high-dimensional features and low computational efficiency.
Furthermore, distinct statistical properties may exhibit diverse degrees of correlation with the sensitivity features of diverse tree species.The ideal feature subset to represent a specific tree species is still difficult to determine.Therefore, it is necessary to select specific image statistical features according to the specific forestry tasks.By analyzing a subset of features that are highly correlated with tree species categories, it is critical to carefully evaluate the relevance of different image statistical properties for specific tree species, identify that can best depict the sensitivity feature subset of distinct tree species to enhance our comprehension of their ecological roles, and develop effective forest management and conservation strategies.The physical traits, comprising the shape of tree trunks and leaves, as well as the odors emitted by bark and flowers, primarily serve as distinguishing characteristics for individual tree species [18].Additionally, Wheeler proposed that the texture features, which are specific to each species of tree, can also be utilized as unique traits for discerning tree species.Wheeler (2011) proposed that texture features exhibited by each wood species may also serve as unique features for identifying tree species [19].
In order to overcome these limitations and accurately obtain the feature subset that can best describe various tree species, this paper considers various image statistical properties and proposes to apply the Deep SVDD model to the sensitive feature analysis of tree species, which can obtain the feature subset that best characterizes individual tree species [20].
Hence, by comprehensively analyzing various image statistical properties of five different tree species, the most suitable feature subset for characterizing each tree species was identified.This will improve our understanding of the uniqueness of the different tree species.This also illustrates that comprehending the sensitivity features of various tree species is essential for forest management and conservation.In this study, we propose a method for the analysis of tree species sensitivity features using machine learning algorithms.First, the image sets of five different tree species were constructed, and various statistical properties were extracted from the tree images.Various feature ranking methods were used to filter out the feature subset that was more relevant to the tree category, and the selected features were used to build an SVM model to accurately identify the number of feature subsets that best characterize the tree category.Combined with the selected feature subset, model optimization was performed using a grid search method.Then, using the specific tree species as a positive sample, Deep SVDD was used to identify the sensitivity characteristics of the specific tree species.Finally, the effectiveness of various tree species feature subsets was verified.The contributions of this paper are summarized as follows: (1) Our method extracted the statistical properties of tree species images and adopted a variety of feature ranking methods to obtain the feature subset with the greatest correlation with the tree species category variable.
(2) This paper employed the Deep SVDD method to obtain the most sensitive feature subset of each type of tree species, so as to realize the unique description of each type of tree species.

Materials
We collected images of five tree species, namely, Eucommia, Metasequoia, Sycamore, Acer truncatum Bunge, and Ginkgo, from the National Botanical Garden of China (39 • 54 N, 116 • 23 E) and Beijing Forestry University (40 • 0 N, 116 • 20 E).The majority of the tree species at the National Botanical Garden of China were Eucommia, Metasequoia, and Acer truncatum Bunge.Sycamore and Ginkgo were the two primary tree species that Beijing Forestry University gathered.
To ensure the consistency of images, all images were captured vertically with a Samsung portable digital camera WB650 in automatic mode at noon each day, and the resolution of the acquired images was 1920 × 1080.For each species, a total of 100 color images were gathered, resulting in a dataset of 500 images.A few typical images are shown in Figure 1.The consistency of the image collection process eliminated the impact of external factors on the dataset, making it an ideal resource for evaluating the unique feature subset of the tree species.
Forests 2023, 14, x FOR PEER REVIEW 3 of 17 the sensitivity characteristics of the specific tree species.Finally, the effectiveness of various tree species feature subsets was verified.The contributions of this paper are summarized as follows: (1) Our method extracted the statistical properties of tree species images and adopted a variety of feature ranking methods to obtain the feature subset with the greatest correlation with the tree species category variable.
(2) This paper employed the Deep SVDD method to obtain the most sensitive feature subset of each type of tree species, so as to realize the unique description of each type of tree species.

Materials
We collected images of five tree species, namely, Eucommia, Metasequoia, Sycamore, Acer truncatum Bunge, and Ginkgo, from the National Botanical Garden of China (39°54′ N, 116°23′ E) and Beijing Forestry University (40°0′ N, 116°20′ E).The majority of the tree species at the National Botanical Garden of China were Eucommia, Metasequoia, and Acer truncatum Bunge.Sycamore and Ginkgo were the two primary tree species that Beijing Forestry University gathered.
To ensure the consistency of images, all images were captured vertically with a Samsung portable digital camera WB650 in automatic mode at noon each day, and the resolution of the acquired images was 1920 × 1080.For each species, a total of 100 color images were gathered, resulting in a dataset of 500 images.A few typical images are shown in Figure 1.The consistency of the image collection process eliminated the impact of external factors on the dataset, making it an ideal resource for evaluating the unique feature subset of the tree species.

Image Statistical Properties
In order to identify the key properties that distinguish each type of tree species, we evaluated them using several statistical properties. (

1) Color property
We took into account the average hue, saturation, and brightness of the image to make the subsequent calculations easier [21].When selecting these features, we specifically analyzed their differences among all tree species.
(2) Texture property To describe the texture condition more intuitively in terms of the co-occurrence matrix, we mainly used energy, contrast, correlation, local homogeneity, and entropy to characterize the gray-level co-occurrence matrix [22].
(3) Shape property The shape of the image was an important aspect of image processing.The invariance of images after translation, scale, and rotation made invariant moments an important feature that was often employed in image analysis and recognition [23].Due to their robust stability and high computational efficiency, these invariant moments are more helpful in

Image Statistical Properties
In order to identify the key properties that distinguish each type of tree species, we evaluated them using several statistical properties.
(1) Color property We took into account the average hue, saturation, and brightness of the image to make the subsequent calculations easier [21].When selecting these features, we specifically analyzed their differences among all tree species.
(2) Texture property To describe the texture condition more intuitively in terms of the co-occurrence matrix, we mainly used energy, contrast, correlation, local homogeneity, and entropy to characterize the gray-level co-occurrence matrix [22].
(3) Shape property The shape of the image was an important aspect of image processing.The invariance of images after translation, scale, and rotation made invariant moments an important feature that was often employed in image analysis and recognition [23].Due to their robust stability and high computational efficiency, these invariant moments are more helpful in distinguishing between different tree species.In this work, the image shape was described by the seven invariant moments.
(4) Power spectrum The structure was the most intriguing aspect of the image.The covariance or correlation between the adjacent pairs of pixels in an image was frequently considered by second-order statistics [24].The distance and direction functions of the image were defined by the autocorrelation function, which quantified the closeness of two-pixel locations.It has been demonstrated that the image power spectrum and the image's autocorrelation function are intimately connected, and the square of the power spectrum's amplitude was regarded as the autocorrelation function's Fourier transform.Therefore, information about the autocorrelation function in the image may be found using the power spectrum of the image.The equation of the average power spectrum model in polar coordinates is shown below.
Each direction's amplitude scale factor was represented by the function A s (θ), while the direction function's frequency exponent was represented by α s (θ).By fitting the model A/ f α to the image power spectrum in four directions (horizontal, tilt, and vertical), α and A were produced. (

5) Weibull distribution coefficients
In image analysis and processing, the Weibull distribution was used as a probability distribution model with two parameters that accurately captured the contrast statistics and edge distribution of an image [25].To calculate the shape and scale parameters of the image Weibull distribution coefficient, we utilized the maximum likelihood estimation (MLE) method and the Weibull fitting method to fit the gradient histogram of the image.The likelihood function was a crucial tool for selecting the appropriate parameters and served as a key indicator of how well the distribution model aligns with the data.The distribution parameters that maximize the log-likelihood function provided the best fit.In this study, fitting the image with the Weibull distribution model extracted more accurate feature information from the image, improving the effectiveness of understanding tree species and facilitating the characterization of their unique sensitivity features.

(6) Mean Subtracted Contrast Normalized coefficients
In addition, we used the Mean Subtracted Contrast Normalized (MSCN) coefficients to describe the brightness distribution in the image [26].According to the calculations, the MSCN coefficient distribution of the forest image was quite similar to the typical Gaussian distribution.As a result, the images can be fitted using the generalized Gaussian density (GGD) distribution, the equation of which is shown below.The first set of features comprised the shape parameter α and variance parameter σ 2 generated through the fitting of the GGD distribution.
is the gamma function.The association information between the pixels was computed using the primary and secondary diagonals, the bottom and right sides of the four directions, and the primary and secondary diagonals.An asymmetric generalized Gaussian (AGGD) distribution was used to approximate the fit to the data.With the aid of a moment matching-based algorithm, the parameters of the AGGD were calculated, which included a total of 16 properties in four directions.The properties of the 18-dimensional brightness distribution for each image were finally obtained.Additionally, the coefficients of the image after discrete cosine transform (DCT) can be fitted using GGD [27].Initially, the image was transformed using DCT, after which it was divided into 5 × 5 equal-sized local image blocks.The discrete cosine transform coefficients of these blocks and the particular partitions within each cosine transform block were fitted using GGD.The formal parameter of the GGD fitting was the first feature of the DCT coefficient.
The following formula represented the local block frequency variation coefficient of the DCT coefficients.
The second feature of the DCT coefficients for the entire image was the average value of all properties.
At the same time, the discrete cosine transform block was divided into three different frequency bands.
In order to obtain the model variance σ 2 n , which served as the average energy E n of the appropriate frequency band n(n = 1, 2, 3), the DCT coefficients for each of these frequency bands were subjected to GGD fitting.Then, the average energy preceding frequency band n was calculated, and the difference between the two was divided by their sum to obtain R n (n = 2, 3).Finally, R 2 and R 3 were averaged over all blocks to obtain the energy subband ratio characteristics.
In order to extract the orientation information of the local image block, we modeled the DCT coefficients of image blocks in three different directions.Then, the variances of the three directions ζ from all the image blocks were combined, producing directionbased properties.
Using the aforementioned four features as statistical properties in the DCT domain, we effectively described the structural information of the image.

(8) Wavelet coefficients
The wavelet transform coefficient can describe the spatial local form and statistical analysis of the image; thus, it can be used to reveal the statistical laws of the image's local structure [28].
Utilizing a controlled pyramid, image decomposition was accomplished by lowering the statistical correlation between the subbands through division normalization.The subband statistics produced by this normalization were more resemblant to a Gaussian model.The wavelet coefficients were extracted using the generalized Gaussian distribution after the image was divided into twelve subbands at two scales and six directions using wavelet decomposition.The generalized Gaussian model was fitted to the subband coefficients to produce the variance and shape parameters, which were used as wavelet statistical properties.A total of 24 features can be derived by fitting the coefficients of each of the twelve subbands.
To obtain suitable distribution coefficients, we employed the 1D generalized Gaussian distribution to fit subband coefficients at various scales in the same direction.As the variance σ 2 in the GGD model lacks practical information, we only utilized the shape parameter γ as a feature value, resulting in seven features that accounted for the onedimensional properties across subbands.
Furthermore, we modeled the correlation between the high-pass (HP) response and band-pass (BP) response of the image using a Gaussian window of size 15 × 15 and σ = 1.5, which enabled us to determine the window structure correlation between each BP Forests 2023, 14, 1057 6 of 17 subband and the HP residual band.This can generate an additional twelve features across the orientations.
Finally, the window structure correlation between all pairs of subbands was described using 15 features.A single image can provide 88-dimensional wavelet statistical features in total.
Table 1 below provides a description of the various picture statistical properties used in this study.These data were utilized to offer a subset of fundamental statistical properties of tree species images.Since certain feature dimensions in this study were excessively high, the approach of adding numbers was used to differentiate the specific dimensions in each type of statistical property.

Feature Correlation Ranking
To identify the most relevant features of each tree species category, each of the aforementioned features underwent various filtering techniques.
(1) Spearman The correlation coefficient was a nonparametric statistical method used to measure the degree of association between two variables, typically their monotonic relationship.Without previous knowledge of the variables or the nature of the connection, the Spearman correlation coefficient can assess the relationship between two random variables based on their respective probability distributions.When dealing with non-normally distributed data, it was frequently utilized to rank the values rather than to use the raw numbers.
A greater absolute value of the Spearman correlation coefficient indicated a stronger correlation between the two variables.The Spearman correlation coefficients ranged from −1 to 1.In general, a very high correlation was observed with an absolute value greater than 0.8.An indication of a positive correlation was a positive value, whereas an indication of a negative correlation was a negative value.The complete positive correlation between the two variables was indicated by a coefficient of 1.
In a dataset with a sample size of n, the variables x i and y i were ranked, and their rank orders were denoted by R i and Q i , respectively.The difference between the rank order of the two variables was represented by (2) mRMR However, when ranking features, the correlation coefficient algorithm only took into account how strongly the features were correlated with the target variables.As a result, the minimum Redundancy Maximum Relevance (mRMR) algorithm was also used in this study to choose the features.In addition to providing inter-feature redundancy limits to remove highly correlated redundant features, the mRMR technique found the maximum correlation between features and target variables [29].The mRMR method had two stages: the first stage filters out features that were significantly linked with the category by using the maximum correlation; different correlation measures, such as mutual information, can be used.The features with the highest mutual information values were mostly chosen for this investigation.However, there may be significant dependencies between the features chosen in the first stage.In order to obtain a collection of mutually independent features that were strongly connected to the category, the second step sought to eliminate the mutually redundant features.
(3) ReliefF The Relief algorithm was a prevalent algorithm used for feature weighting, which assessed feature weights based on differences between similar samples to determine their influence on the classification results.When dealing with various categorization issues, missing information, and noise, this algorithm had certain limitations.The Relief algorithm may be inefficient in processing data that contain missing information.Additionally, the presence of noise or outliers in a dataset can negatively impact the algorithm's performance, potentially leading to inaccurate results.It is worth noting that the Relief algorithm is generally only capable of handling two types of data [30].
To address these shortcomings, Kononenko et al. proposed an improved algorithm called ReliefF, which treated multi-classification problems as a one-to-many situation [31].The ReliefF algorithm finds the nearest neighbors of each type of sample and synthesizes a computational analysis to effectively extract feature information related to tree categories, providing guidance for subsequent modeling or prediction.A larger final weight value indicates that the feature is more important.The weight-updated formula for the feature in ReliefF is shown in the following equation.
The category labels of the features and tree species were calculated utilizing the ReliefF method with a closest neighbor K value of 5, resulting in feature weights that were sorted sequentially.Additionally, it should be noted that the rank of a feature decreased as its weight increased throughout the feature selection process.

Feature Importance
A critical concept in predictive modeling for data analysis was feature importance ranking, which revealed the correlation between features and the target variable and offers insights into the dataset and model to improve the efficiency and efficacy of the model.However, given that datasets usually contain a vast number of features, determining the most important ones became a significant challenge.
Features with higher scores indicated greater relevance for the target, and prediction models based on the training set calculated the feature importance scores to represent the value of various features in predicting the target variable.The relevance of features may be quantified using distance, which measured the ability of features to differentiate samples from different categories, and information score, which compared the entropy variation to determine the quantity of information contained in the features [32].
The information score (IS) is a method of calculating the feature importance by utilizing the entropy value of the data and expressing the feature importance through entropy variation.On the other hand, the intra-class distance (ICD) is a metric that measures the distance between each point within a class and its center; smaller distances indicate a better state of the data.Although both IS and ICD can independently evaluate the relevance of a feature for a given tree species, they may produce different rankings for the same feature.
To aggregate the feature ranking results and avoid errors, this study implemented an averaging strategy that involved taking the average of the ranks produced by the two measures.

Deep SVDD
The model of Deep SVDD utilizes neural network training to minimize the division hypersphere of the feature space in the sample.By calculating the distance between the test sample point and the sphere's center, the model was able to assess whether a sample point is anomalous.The feature points of the sample points were mapped by the neural network into the hypersphere with the least volume.
X ⊆ R d stood for the input sample space, F ⊆ R p represented the output, φ(•; W) : X → F represented the feature mapping function of the neural network, and W = W 1 , . . ., W L represented the corresponding parameters.The feature representation of a sample x ∈ X was φ(x; W).In the Deep SVDD, the primary goal was to minimize the volume of the hypersphere that encloses the output feature space of the sample points.The center c and radius R represented the parameters of the hypersphere.This was achieved through neural network training and the soft-boundary optimization function defined for a set of training sample points D n = {x 1 , . . . ,x n }.
When dealing with a test sample point x ∈ X, the anomaly score was computed by measuring the distance between the sample point and the center of the hypersphere in the output space.This computation produced an expression that can be used to determine the anomaly score.
where W * represented the weight of the trained model.

Validation
The training dataset was constructed by randomly sampling approximately 70% of the original dataset, whereas the remaining 30% of the samples were allocated for data validation.To evaluate and assess the performance of the feature subset, this experiment calculated the overall accuracy, precision, recall, and F1-score of the model by using the confusion matrix derived from the model's predictions on the test samples.
Accuracy was the proportion of all samples with positive predictions in which the labels were the same as the true samples.This value indicated the confidence of the model for correct predictions.

ACC =
TP + TN TP + TN + FP + FN (10) Precision was the probability that the predicted value of all the results obtained from the prediction of the sample was positive class, which was actually positive.
Recall meant the proportion of all true positive samples that were correctly predicted by the model to be positive.
Precision and recall were calculated around the positive samples, so the F1-score was used to reflect the performance of the model in aggregate.

Data Visualization
Selecting the appropriate visualization tool was crucial for performing an initial study of potential variations in statistical data among different tree species.In this regard, for each statistical property dimension of the five categories of tree species, we utilized violin plots to visualize the data distribution.An example of a violin plot for certain statistical properties is depicted in Figure 2.
Precision and recall were calculated around the positive samples, so the F1-score was used to reflect the performance of the model in aggregate.

Data Visualization
Selecting the appropriate visualization tool was crucial for performing an initial study of potential variations in statistical data among different tree species.In this regard, for each statistical property dimension of the five categories of tree species, we utilized violin plots to visualize the data distribution.An example of a violin plot for certain statistical properties is depicted in Figure 2. From Figure 2, it is evident that there are substantial differences in the distribution of different tree species for particular feature values.By observing the violin plots of the statistical properties constructed related to the tree species categories, it can be shown that the majority of the statistical properties have a positively skewed distribution resembling From Figure 2, it is evident that there are substantial differences in the distribution of different tree species for particular feature values.By observing the violin plots of the statistical properties constructed related to the tree species categories, it can be shown that the majority of the statistical properties have a positively skewed distribution resembling the normal distribution.While Sycamore and Eucommia had more concentrated data distributions, Eucommia had a somewhat skewed distribution, as seen by shape_property.1.By further comparing distributions of five different tree species, it was found that the values of each quantile of Ginkgo were higher than those of other tree species in Texture_property.4.In addition, in the mean_s, the median of Metasequoia was slightly higher than that of the other tree species.Additionally, Sycamore's distribution values in the Weibull_distribution.1 were all lower than those of the other tree species.It should be noted that, in the presentation of the kernel density function for this value, all tree species categories showed similar distributions.In the Wavelet_coefficients.23, it was found that the Sycamore and Metasequoia exhibited a distinct bimodal distribution, while the difference between the maximum and minimum values of Metasequoia species was in the extreme position among the five categories.

Feature Ranking
After conducting the feature correlation analysis, features with a correlation coefficient greater than 0.8 were eliminated.Then, 11 features were retained to improve the model's interpretability.These features were then used to identify the sensitivity feature subsets of the different tree species.To determine the feature ranking, we utilized a variety of feature filtering methods, such as Spearman, mRMR, and ReliefF.We computed the average of the feature rankings acquired from the three different feature filtering methods to attain a comprehensive feature ranking.The feature level shows the relationship between the feature and the tree species.To obtain a comprehensive feature ranking, we averaged the feature sequences obtained using the three different feature filtering methods.A lower feature level indicates a higher correlation between the feature and the tree species.Table 2 displays the feature rankings derived from the aforementioned algorithms.Texture_property.4 showed a strong correlation with the tree species category when computed using the mRMR and ReliefF algorithms, but its correlation was poor when computed using the correlation coefficient method.After combining the ranks obtained from all three methods, Weibull_distribution.1 exhibited the strongest correlation with the tree species category, while MSCN coefficients.11and Wavelet coefficients.23 had lower correlations.

Accuracy Analysis
Eleven variables were selected as input variables for the model to determine the ideal number of tree species-sensitive features.The SVM was chosen to filter the number of features that strongly influenced the tree species category.Only the ranking of features obtained after thorough comparison was used in the final confirmation of the optimal number of features, and as the number of selected features steadily rose, each evaluation index was computed in sequence.The relationship between the number of features and the classification evaluation metric was examined using the SVM classifier.The entire dataset was split into two parts, where 70% of the data was used as the training set and the remaining as the test set.Evaluation metrics, including accuracy, precision, recall, and F1-score, were used to evaluate the model, and the results are presented in Figure 3.

Accuracy Analysis
Eleven variables were selected as input variables for the model to determine the idea number of tree species-sensitive features.The SVM was chosen to filter the number o features that strongly influenced the tree species category.Only the ranking of feature obtained after thorough comparison was used in the final confirmation of the optima number of features, and as the number of selected features steadily rose, each evaluation index was computed in sequence.The relationship between the number of features and the classification evaluation metric was examined using the SVM classifier.The entire da taset was split into two parts, where 70% of the data was used as the training set and th remaining as the test set.Evaluation metrics, including accuracy, precision, recall, and F1 score, were used to evaluate the model, and the results are presented in Figure 3.In Figure 3, it can be observed that an increase in the number of features led to a ris in the overall accuracy of all evaluation indices.When six features were present, the F1 score reached its maximum value of 0.662.However, beyond a feature index of eight, th In Figure 3, it can be observed that an increase in the number of features led to a rise in the overall accuracy of all evaluation indices.When six features were present, the F1-score reached its maximum value of 0.662.However, beyond a feature index of eight, the accuracy was found to be inversely associated with the number of features, and the total accuracy, precision, and recall all reached their optimal values at this point.
To optimize the hyperparameters of SVM, grid search methods were employed, where the penalty function and kernel function's parameter ranges were varied between 2 −5 and 2 5 at a step size of 0.1.The K-fold cross-validation method was used for the model validation, splitting the dataset into five groups as k was set to five in this experiment.During each of the five algorithm iterations, the model was trained on four datasets, while one of these five sets was used as a test set to check the accuracy of the model's classification.Finally, the overall performance of the model was assessed using the average of evaluation metrics calculated from the test set.Figure 4 illustrates the optimization process of the grid search algorithm for parameters in the image statistical feature subset of different tree species.Upon exploring all possible combinations of hyperparameters, the best classification accuracy achieved on the dataset was 0.7, with an ideal parameter set of (2.83, 0.031).
uation metrics calculated from the test set.Figure 4 illustrates the optimization process of the grid search algorithm for parameters in the image statistical feature subset of different tree species.Upon exploring all possible combinations of hyperparameters, the best classification accuracy achieved on the dataset was 0.7, with an ideal parameter set of (2.83, 0.031).

The Sensitivity Features Analysis
Table 3 displays the average of the two feature order indicators for each dimension feature.Moreover, the table highlights the significant differences in feature sensitivity across various tree species categories.However, Weibull_distribution.1 consistently ranked high across all tree species categories.

The Sensitivity Features Analysis
Table 3 displays the average of the two feature order indicators for each dimension feature.Moreover, the table highlights the significant differences in feature sensitivity across various tree species categories.However, Weibull_distribution.1 consistently ranked high across all tree species categories.When calculating the ensitivity features of a specific category of five types of tree species images, 80 samples from that category were chosen as the training set.The test set was created by combining the remaining 20 samples with 20 samples from each category of tree species that were chosen at random.The number of features gradually increased in the order from front to back as the input of the model of Deep SVDD for training.
The F1-score was utilized to determine the optimal number of sensitive features for each tree species category.The subset of sensitive features required for the category was obtained when the F1-score reached its overall optimal value or attained a value of 0.7.
Figure 5 illustrates that as the number of features increased, the accuracy of the final indexes for Eucommia and Metasequoia reached 0.8.For the Sycamore, the F1-score peaked at 0.82 when the number of features reached six.The evaluation indicators for the two species of Acer truncatum Bunge and Ginkgo ingot displayed an initial increase, followed by a decline, and then increased again, with the overall evaluation index tending to increase as the number of features increased.At the maximum number of features, the F1-score of these two tree species reached the maximum values of 0.679 and 0.655, respectively.peaked at 0.82 when the number of features reached six.The evaluation indicators fo two species of Acer truncatum Bunge and Ginkgo ingot displayed an initial increase lowed by a decline, and then increased again, with the overall evaluation index ten to increase as the number of features increased.At the maximum number of features F1-score of these two tree species reached the maximum values of 0.679 and 0.655, re tively.Hence, as the number of features increased, it became possible to obtain the subs sensitive features for various tree species categories.The testing of five categories o species, including Eucommia, Metasequoia, Sycamore, Acer truncatum Bunge, Ginkgo, resulted in obtaining 3, 6, 6, 8, and 8 sensitive feature sets, respectively, the determining the set of sensitive features for each tree species.
Among these sets of sensitive features, the optimal feature combination for ch terizing Eucommia comprised Weibull_distribution.  Hence, as the number of features increased, it became possible to obtain the subset of sensitive features for various tree species categories.The testing of five categories of tree species, including Eucommia, Metasequoia, Sycamore, Acer truncatum Bunge, and Ginkgo, resulted in obtaining 3, 6, 6, 8, and 8 sensitive feature sets, respectively, thereby determining the set of sensitive features for each tree species.

Validation
To validate the sensitive feature subsets of different tree species, the accuracy of the results was analyzed and verified.We only selected the corresponding feature value and regarded the category of this tree species as a positive sample, with other samples considered negative to verify the validity of a particular tree species' sensitive feature subset.Five iterations of the algorithm were performed to ensure accuracy.The results, along with the various evaluation indicators, are presented in Table 4.It should be noted that the classification precision for Sycamore was 0.771, and all five tree species obtained F1-score exceeding 0.7.This demonstrates that the sensitive feature subsets of different tree species can effectively capture their unique features.

Conclusions
At present, image statistical properties are widely used and are powerful descriptive features in image processing [33].The image statistical properties of tree species are essentially digital and quantitative descriptions of tree species characteristics, which are highly objective and comparable [34].The extraction and analysis of these features can provide important data support and reference for tree species classification, ecological research, growth environment analysis, and ecological protection [35].
However, it is important to note that not all image statistical properties are equally correlated with the tree species.For example, certain features may be more relevant to problems such as pest detection or disease identification [36].Problems such as tree species identification and classification can be realized by using features that are highly correlated with the tree species [37].Furthermore, the accuracy of the 3D reconstructions of forest scenes can also be improved by using features that are less linked to tree species [38].To identify the most relevant features for tree species variables and understand the sensitive feature subsets among diverse tree species, we employed various image statistical properties to quantify the tree images.Additionally, we analyzed the correlation between the image statistical properties and tree species categories.Eventually, we identified a subset of sensitive features for each tree species and verified the feasibility of our method.These experimental results emphasize the importance of carefully selecting image features for specific forestry tasks and demonstrate the potential of image-based methods for tree species analysis.
The SVM is a widely utilized classification algorithm in the field of forestry [39].The performance of the SVM model in representing the tree species was assessed by calculating various evaluation metrics.It should be noted that not all features have a high correlation with the tree species variable, meaning that having more features does not necessarily lead to higher classification accuracy for tree species [40].Feature selection is an essential part of feature engineering, which aims to choose the most relevant features from all the features [41].The filter method is commonly used to identify the most significant feature subset that is highly correlated with the target label.This method is favored for its simplicity and broad applicability across diverse domains [42].When classifying tree species using a limited number of features, enhancing the additional features can significantly increase the distinction between different tree species.Nevertheless, beyond a certain threshold of the number of features, adding more features does not significantly improve tree species classification and recognition [43].Once the saturation point of the number of features has been reached, adding more features will lead to data redundancy, which can negatively affect the model performance and reduce its effectiveness.The results demonstrate that the correlation of image statistical properties with tree species variables does not always increase with the number of features.By combining the filtering method with SVM, we identified eight feature subsets that have the strongest correlation with the tree species categories.
Furthermore, feature importance measures are commonly employed for analyzing feature ranking problems and assessing the global sensitivity of classification models for each feature independently [44].At the same time, feature evaluation and selection techniques based on SVDD can also be applied to feature selection in one-class classification problems [45].However, the use of SVDD to analyze tree species sensitivity feature subsets is a novel research topic.
In this study, we observed differences in the sensitivity feature subsets among various tree species.The F1-score of all tree species was higher than 0.7 when using the sensitive feature subset of each tree species to verify its effectiveness.The experimental results demonstrate that our proposed method is effective in identifying sensitive feature subsets in tree species classification.This method can provide valuable insights into the development of more efficient methods for identifying tree species, ultimately contributing to the management and conservation of forest ecosystems.Nonetheless, our dataset was limited to certain tree species and geographic locations.While providing valuable insights and results, further research is required to explore the model's performance on datasets from other regions with different tree species and variations in image acquisition conditions.Understanding the relationship between image statistical properties and other tree species can facilitate the model's applicability to broader scenarios and enhance the evaluation metrics of algorithmic classification.Future studies could focus on expanding the dataset to include a more comprehensive range of tree species and geographic locations to enhance the model's robustness.Additionally, exploring other factors that can affect the model's performance, such as environmental conditions and image quality, could open up new avenues of research to further optimize the model for practical applications under real-world conditions.

Figure 1 .
Figure 1.Typical picture in the tree species dataset.

Figure 1 .
Figure 1.Typical picture in the tree species dataset.

Figure 2 .
Figure 2. Violin plots of the tree species.

Figure 3 .
Figure 3.Comparison of accuracy of selected feature subsets using SVM classifier by combinin multiple feature selection ranking methods.

Figure 3 .
Figure 3.Comparison of accuracy of selected feature subsets using SVM classifier by combining multiple feature selection ranking methods.

Figure 4 .
Figure 4. Optimization diagram of the grid search algorithm parameters.

Figure 4 .
Figure 4. Optimization diagram of the grid search algorithm parameters.

Figure 5 .
Figure 5. Relationship between the number of features and F1-score.

Figure 5 .
Figure 5. Relationship between the number of features and F1-score.

Table 1 .
A summary of the image's statistical properties.

Table 2 .
Results of applying average feature rank to the tree species dataset.

Table 3 .
Results of applying average feature rank to the tree species dataset.

Table 3 .
Results of applying average feature rank to the tree species dataset.

Table 4 .
Multiple evaluation metrics for tree species.