Monitoring Fine-Scale Forest Health Using Unmanned Aerial Monitoring Fine-Scale Forest Health Using Unmanned Aerial Systems (UAS) Multispectral Models Systems (UAS) Multispectral Models

: Forest disturbances—driven by pests, pathogens, and discrete events—have led to billions of dollars in lost ecosystem services and management costs. To understand the patterns and severity of these stressors across complex landscapes, there must be an increase in reliable data at scales compatible with management actions. Unmanned aerial systems (UAS or UAV) offer a capable platform for collecting local scale (e.g., individual tree) forestry data. In this study, we evaluate the capability of UAS multispectral imagery and freely available National Agricultural Imagery Program (NAIP) imagery for differentiating coniferous healthy, coniferous stressed, deciduous healthy, deciduous stressed, and degraded individual trees throughout a complex, mixed-species forests. These methods are ﬁrst compared to assessments of crown vigor in the ﬁeld, to evaluate the potential in supplementing this resource intensive practice. This investigation uses the random forest and support vector machine (SVM) machine learning algorithms to classify the imagery into the ﬁve forest health classes. Using the random forest classiﬁer, the UAS imagery correctly classiﬁed ﬁve forest Health classes with an overall accuracy of 65.43%. Using similar methods, the high-resolution airborne NAIP imagery achieved an overall accuracy of 50.50% for the ﬁve health classes, a reduction of 14.93%. When these classes were generalized to healthy, stressed, and degraded trees, the accuracy improved to 71.19%, using UAS imagery, and 70.62%, using airborne imagery. Further analysis into the precise calibration of UAS multispectral imagery, a reﬁnement of image segmentation methods, and the fusion of these data with more widely distributed remotely sensed imagery would further enhance the potential of these methods to more effectively and efﬁciently collect forest health information from the UAS instead of using ﬁeld methods.


Introduction
Forest disturbances, coupled with invasions by foreign pests and pathogens, have dramatically altered vegetation systems. These discrete events transform physical structure, ecosystem processes, and resource allocations which play a significant role at local and global scales and across both natural and developed environments [1][2][3][4]. Examples of prevalent forest disturbance include fires, flooding, windstorms, droughts, overharvesting, pollution, fragmentation, and biological invasions. Invasions by insects and pathogens threaten the stability of forest ecosystems, events that are projected to increase [5,6]. Private landowners and local governments most heavily endure the degradation and ecosystem change caused by these biological invasions [5,7,8]. In conjunction with distinct disturbance events, continuous stress from anthropogenic activities has had a measured impact [9]. Managing forests for peak growth requires not only a combination of nutrients, light, temperature, and moisture but also the absence or diminished presence of threats and invasions [9][10][11]. Reconciling forest disturbances and stress requires understanding where it occurs and what influences it may have at several spatial and temporal scales. Despite ecosystems. A comprehensive evaluation of forest health, including multiple observations or measurements, is needed for effective monitoring [2,8,35].
To observe, or measure, a reduction in forest health using remote sensing requires choosing a well-fitting indicator [25,[36][37][38]. Choosing such an indicator is often one of the first steps in a forest health assessment [16]. At both the individual tree level and at the level of forest stands or landscapes these indicators measure conditions, or the changes in them using a range of variables. These variables include crown vigor [6,39,40]; structural characteristics such as tree height, growing stock, or crown size [25]; phenology; water content; defoliation [18,41]; canopy discoloration; and fragmentation [9,42]. Two prevailing techniques exist for conducting these assessments of forest health using modern, high-resolution, imagery: (1) aerial surveys or (2) digital image classification.
Aerial visual surveys using piloted aircraft provide excellent scales of observation, with potentially highly accurate results. The methods for conducting these surveys are well-defined and broadly adopted, allowing trained personnel to detect both tree species and disturbance type (e.g., disease or defoliation cause) even among complex forests [3,17]. Still, these surveys are restricted due to their: cost, inability to fly on-demand, insufficient temporal frequency for observing all types of disturbance, and limitation of only detecting areas that already show signs of invasion or stress. Close range digital remote sensing and satellite imagery classification allows users to precisely monitor long-term stress and change using indicators [8,25]. Digital imagery with modern hardware often assimilates the use of signal theory and increased spectral dimensionality for obtaining information from measured reflectance [43]. A primary example of this is chlorophyll fluorescence measurements, which assess the photosynthetic efficiency, or variation in it (stress), to determine vegetation status or health [8,34]. When a plant becomes stressed, due to the impacts of some stressors, it is said to be 'chlorotic' which results in a reduction in photosynthetic activity and is marked by a shift towards greater amounts of green and red reflectance [44,45]. Both healthy and stressed leaves can be identified based on the internal and external structures of the leaf (e.g., chlorophylls and xanthophyll) and their responses to electromagnetic energy [44,46]. From small handheld devices to multimillion-dollar platforms, sensors are being developed and applied that can detect stress or changes in photosynthetic efficiency (i.e., metabolism) before visible indications are available [8,47]. These remote sensing spectral responses are tested against laboratory analyses to distinguish between true and observed reflectance (i.e., defining necessary radiometric corrections) [43,48,49]. Once the true reflectance from a given sensor is defined, statistical relationships between spectral response and various biotic traits can be empirically modeled [8,50]. For example, many spectral band indices have been developed which can be used to interpret changes in vegetation status or condition [6,41,50]. The normalized difference vegetation index (NDVI), including modified versions, and forms of visible vegetative indices (VVI) represent two of the most readily applied methods [51][52][53][54]. However, in the current literature, there are countless vegetation indices used for both general and precise purposes [9,44].
More advanced sensors, such as Goddard's LiDAR Hyperspectral and Thermal Imager (G-LiHT), bring together hyperspectral imaging and 3D laser scanner reconstructions to form fusion datasets. G-LiHT and comparable sensors are able to map vegetation communities, invasive species presences, natural disturbances, and carbon cycles [8,12,55,56]. Multisensor data fusion has become more prominent in recent decades. However, challengessuch as spectral intercalibration, temporal discontinuity, and positional misregistrationmust be managed when adopting these methods [57][58][59]. Like aerial visual surveys, these digitally classified remotely sensed images can be still limited by temporal infrequencies, inflexible deployment conditions, cost, and cloud coverage [57,58,60]. Therefore, the question remains of how to best harmonize these evolving technologies with operational feasibility.
Unmanned aerial systems (UAS) have become a noteworthy platform for bringing geospatial sciences and technologies into the hands of more diverse stakeholders. Although UAS has had a long history of military development, consumer market demands and concurrent technological innovations have made this platform both economic and adaptive [61,62]. Several studies have used normal color or modified normal color consumer-grade cameras onboard UAS to measure vegetation biophysical properties with high precision [50,[63][64][65]. Other studies have applied UAS for estimating attributes of individual trees and forest stands [66][67][68][69]. These efforts directly assist the need for largescale (i.e., individual tree or management unit size) data which can be used for disturbance monitoring and decision making [19,[70][71][72]. This study further defines a niche for UAS forest health assessments, between that of advanced data fusion techniques and more limited yet operational aerial surveys. This research provides a means for large-scale (local) land managers to have a more complete understanding of their forests, by supplementing in situ surveying. The evaluation presented here evaluates this application through a comparison of UAS multispectral image analysis to an established, multi-factor, field-based assessment of crown vigor [6,30]. By providing information on the presence and abundance of stressed or degraded trees, using UAS, instead of relying on methods most commonly applied through visual assessments in the field, forest managers can more quickly react to lowered resource availability or diminished ecosystem function [40]. For this reason, we investigated the ability to classify coniferous and deciduous tree health classes, instead of targeting a specific disturbance event. To accomplish this, we evaluated the ability of simple multispectral sensors onboard UAS for distinguishing healthy, stressed, and degraded trees in complex, mixed-species, forests. Conducting this study throughout such natural environments provides direct relevance to many regional land managers, which is not commonly found in similar investigations, who face the complexity of managing mixed-species forests [14,65,70,73]. Specifically, our objectives were: 1. Determine the capability of UAS for classifying forest health at the individual tree level. 2. Compare the results of forest health classification using UAS to high-resolution, multispectral, airborne imagery.

Study Areas
Four woodland properties, managed by the University of New Hampshire (UNH) were employed in this research. These properties included: Kingman Farm, Thompson Farm, College Woods, and Moore Fields ( Figure 1) and represent a total of 304.1 hectares (ha) of forests located near the main UNH campus. These study sites were chosen due to the availability of previous forest inventory records and for having a known presence of forest disturbances (e.g., WPND, HWA, EAB, and beech bark disease) [74,75].

Assessing Forest Health: Field and Photo
Field-based sampling was conducted to provide reference data for each forest health class. At each study area, preexisting continuous forest inventory (CFI) plots were visited to locate a variety coniferous and deciduous species [75,76]. These species included: Eastern white pine (Pinus strobus), eastern hemlock (Tsuga canadensis), red pine (Pinus resinosa), American beech (Fagus grandifolia), red maple (Acer rubrum), white ash (Fraxinus americana), and northern red oak (Quercus rubra). Each individual tree was positionally located using a high-precision EOS Arrow 200 RTK GPS [77]. The positional error, as reported by the device during sampling, ranged between 0.48 m and 3.19 m. Additional trees were located for each health class while traversing the miles of trials distributed throughout each of the properties. All field measurements were made during June and July of 2020 and 2021. To assess the health of each sampled tree, a team of two researchers used visual guides of crown vigor and degradation [6,30]. These visual charts and classifications are based on Pontius and Hallett [30], and supplemental practices suggested in Broders et al. [17] and Innes [29]. This procedure was adopted due to the ease of implementation and available training. Using these charts, data on fine twig dieback, leaf discoloration, leaf defoliation, crown vigor, crown transparency, and crown light exposure (see Pontius and Hallett [30], or Hallett and Hallett [78], for definitions) were entered into the Healthy Trees Healthy Cities app [33]. This app then summarized the full suite of tree health attributes, using standardized variables (Z-scores) which were calculated using the mean and standard deviation of regional, species-specific, observations for each attribute [30,78,79]. For the final step, this app translated this comprehensive, species-specific, Z-scores for each tree into a 10-part, numeric, classification system, with lower values representing healthier trees [30,78].
Remote Sens. 2021, 13, x FOR PEER REVIEW 5 of 23 Figure 1. Four woodland properties evaluated during the assessment of forest health. Each property is shown using the multispectral (false color composite) orthoimagery generated from the unmanned aerial system (UAS) imagery.

Assessing Forest Health: Field and Photo
Field-based sampling was conducted to provide reference data for each forest health class. At each study area, preexisting continuous forest inventory (CFI) plots were visited to locate a variety coniferous and deciduous species [75,76]. These species included: Eastern white pine (Pinus strobus), eastern hemlock (Tsuga canadensis), red pine (Pinus resinosa), American beech (Fagus grandifolia), red maple (Acer rubrum), white ash (Fraxinus americana), and northern red oak (Quercus rubra). Each individual tree was positionally located using a high-precision EOS Arrow 200 RTK GPS [77]. The positional error, as reported by the device during sampling, ranged between 0.48 m and 3.19 m. Additional trees were located for each health class while traversing the miles of trials distributed throughout each of the properties. All field measurements were made during June and July of 2020 and 2021. To assess the health of each sampled tree, a team of two researchers used visual guides of crown vigor and degradation [6,30]. These visual charts and classifications are based on Pontius and Hallett [30], and supplemental practices suggested in Broders et al. [17] and Innes [29]. This procedure was adopted due to the ease of implementation and available training. Using these charts, data on fine twig dieback, leaf discoloration, leaf defoliation, crown vigor, crown transparency, and crown light exposure (see Pontius and For this analysis, we collapsed this 10-part classification system into five distinct forest health classes: • Coniferous (C)-Healthy coniferous trees (e.g., eastern white pine or eastern hemlock) identified as having minimal or no signs of stress, which are calculated using the stress index as classes 1, 2, or 3. • Deciduous (D)-Healthy deciduous trees (e.g., American beech, white ash, or Northern red oak) identified as having minimal or no signs of stress, which are calculated using the stress index as classes 1, 2, or 3. • Coniferous Stressed (CS)-Stressed coniferous trees, displaying moderate or severe reductions in crown vigor, which are calculated using the stress index as classes 4 through 9.
• Deciduous Stressed (DS)-Stressed deciduous trees, displaying moderate or severe reductions in crown vigor, which are calculated using the stress index as classes 4 through 9. • Degraded/Dead (Snag)-Coniferous or deciduous trees identified as stress class 10 (dead) which represent the most degraded of each health attribute.
A minimum of 20 samples for each of these five classes were collected during our fieldinventory. Using these field samples, interpretation guides for each class were established (see Appendix A, Table A1). These guides were then used by a trained forest technician, in addition to ultra-high-resolution, multispectral UAS imagery, to photo interpret additional reference samples. These reference samples were generated from a collection of trees previously measured for structural attributes during the 2019, 2020, and 2021 field seasons (see Fraser and Congalton,[76]) as well as trees located within degraded stands that were visited during the initial field sampling for this study. Photo interpretation was conducted to provide a minimum of 70 samples for each forest health class providing for a more evenly distributed sample throughout the study areas.

Airborne Imagery
To examine the performance of digitally classifying these five forest health classes using freely available, high-resolution, remotely sensed imagery the first analysis was conducted using 2018 National Agriculture Imagery Program (NAIP) imagery. These images were collected between 6 August and 16 October 2018, at a 60-cm spatial resolution, with four spectral bands (blue, green, red, and near infrared (NIR)) [80]. To provide an evaluation of individual trees, these images were segmented using a multiresolution segmentation algorithm within eCognition v9.1 (Trimble, Munich, Germany). The segmentation parameters, as refined in a previous study [81] were: Scale 10, Shape 0.2, and Compactness 0.5. These parameters provided an over segmented result, which was necessary for digitally classifying individual trees. For each image object, 30 object level features were calculated including: spectral, textural, and geometric attributes, as well as three spectral indices (NDVI, NGRDI, and the Greenness Index). These spectral indices were selected due to their given association with plant stress [46,50,53,82]. The equations for NDVI, VVI, NGRDI, and the Greenness Index are given below (Equation (1) through (4)).

UAS Imagery
UAS imagery was collected using a combination of two aircraft, the senseFly eBee X and its predecessor the eBee Plus ((senseFly, Wichita, KS, USA) [83,84]. To obtain natural color imagery, the eBee Plus was operated with its associated sensor optimized for Remote Sens. 2021, 13, 4873 7 of 22 drone applications (SODA.) while the eBee X utilized the senseFly Aeria X sensor [85,86]. These sensors provided the photogrammetric basis for the marker-controlled watershed segmentation (MCWS) described in the next section as well as uncalibrated blue, green, and red spectral bands. Multispectral UAS imagery was collected using the Parrot Sequoia+. This five-lens sensor system is comprised of a natural color sensor (not used in this study), as well as independent green (550 ± 40 nm), red (660 ± 40 nm), NIR (790 ± 40 nm), and red edge (735 ± 10 nm) sensors [87]. All missions were conducted using the eMotion flight management software [88]. The flight parameters for all missions consisted of 85% forward overlap between images, 90% side overlap, consistent sun-angles and cloud exposures, and flying heights of 121.92 m (400 ft) above the ground [62,89,90]. All UAS missions were flown to collect leaf-on (summer) imagery, throughout July and September of 2019 and 2020. Prior to missions conducted using the Parrot Sequoia+ sensor, a radiometric calibration target was used to adjust the camera reflectance to absolute measurements [87]. During post-processing, individual image locations were positionally corrected using the National Oceanic and Atmospheric Administrations (NOAA) Continuously Operating Reference Stations (CORS) and the aircraft's flight logs [91]. The positionally corrected images were then brought into Agisoft MetaShape v1.5.5. (Agisoft LLC, St. Petersburg, Russia) for Structure from Motion Multi-View Stereo (SfM-MVS or SfM) modelling. For each study area, a set of both natural color and multispectral images were processed using the provided SfM workflow within this software. The "High Accuracy" image alignment option was selected, then the "Ultra High" setting for each of the remaining modelling steps [62,81,92]. An ultra-high-resolution digital elevation model (DEM), using the 'mild' point cloud filtering selection, was generated from the natural color imagery to support the segmentation process. Two orthomosaics (i.e., orthoimages) were produced for each property; one from each of the natural color and multispectral workflows.
The UAS imagery was segmented using a MCWS technique outlined in Gu et al. [92], [76,93]. First, a canopy height model (CHM) for each of the four study areas was created by subtracting a 2 m New Hampshire LIDAR bare earth model from the UAS DEMs [94]. A Gaussian (low pass) filter was then applied to these CHMs to remove residual noise in the data [92,93,95]. To establish the individual treetops (i.e., 'markers'), a fixed, circular, window size of 4.5 m was used to identify the local maxima. This window size was found to provide a more accurate single tree delineation in previous studies [76,92,96]. An object detection rate (ODR) and segmentation quality rate (QR) for these data and study areas were published in a previous study, Fraser and Congalton, [76]. Following the individual tree detection and delineation (ITDD) process, we created a composite of the natural color and multispectral UAS imagery for each study area. A nearest neighbor raster resampling tool, within ArcGIS Pro v2.8.0 (Redlands, CA, USA), was used to resample the higher spatial resolution natural color imagery to match the respective study areas multispectral imagery [97][98][99]. This resampling ensured we retained spatial data consistency during the classification process [59,96,100]. These composite images were then used to generate 36 image object features in eCognition (see Appendix A, Table A2).

Forest Health Accuracy Assessment
For the forest health assessment of both the NAIP and UAS imagery, the final check of the reference trees was conducted using photo interpretation and manual (on-screen) editing. Points that could not be matched to corresponding species (i.e., nearby image objects) in either set of imagery were removed. The final sample size for each forest health class for each set of imagery are in Table 1. To quantify the accuracy of classifying for health classes using each source of imagery, we adopted thematic map accuracy assessment error matrices [101]. A number of accuracy assessment (i.e., training and validation data splitting methods) and classification techniques were applied to analyze the results generated from the UAS and NAIP imagery. All digital classifications were performed in Python, using the Scikit-learn package [102]. For the NAIP imagery, all tests were performed using a random forest (RF) supervised classification algorithm [76,103,104]. For the UAS imagery, in addition to using the RF classification algorithm, the support vector machine (SVM) algorithm was also employed [105,106]. This secondary algorithm was included due to the often case-specific superior classification performance found between these two techniques [81,104,107]. The RF classification was performed using a set of 500 trees, and with the Gini index selected as the evaluator for decision tree splits. The SVM algorithm was implemented using a linear kernel. Additional information on the full list of tuning parameters for each of these classifications can be found in Fraser and Congalton [76]. When using the RF classification algorithm, the following analyses were applied: (1) a standard cross-validation with a split of 55% training data and 45% validation data; (2) this same approach with a 50% training and validation data split; (3) splitting the training and validation data 55%/45% but with the removal of the least important image features (i.e., feature reduction); (4) performing the validation using an out-of-bag (OOB) permutation; (5) classifying coniferous and deciduous tree health classes independently; and (6) collapsing the forest health classes into only 'healthy' (a combination of coniferous and deciduous trees), 'stressed', and 'degraded'. Two additional tests were applied to the UAS image classification, using the RF algorithm, to investigate the influence of the redundant image bands included when making a composite of the natural color and multispectral imagery. Each evaluation was performed a minimum of 10 times, so that an average overall accuracy could be produced. For both the NAIP and UAS imagery, a mean decrease in impurity (MDI) test was used to quantify the importance of individual spectral, geometric, and textural image features. The SVM classifier was applied only to the UAS imagery. This classification included a standard cross-validation, with a split of 55% training and 45% validation data (similar to the first RF classification analysis above). This SVM classification was also completed 10 times, so that an average overall classification accuracy could be compared to the RF classification results.

Airborne Imagery
The first assessment of forest health using digitally classified thematic layers was implemented using the freely available NAIP imagery. The individual classification results from each method and averaged (10 trials) overall accuracies can be seen in Table 2. In this table, we see that the highest overall accuracy, when including all five classes, was achieved using a 55%/45% training and validation sample split and the removal of the least important image features (i.e., feature reduction) (Figure 2). The out-of-bag (OOB) accuracy for this same method resulted in a 10.7% lower overall accuracy. When the forest health classes were generalized to only 'healthy', 'stressed', and 'degraded', the overall accuracy reached 70.62%. This average accuracy is also similarly achieved when classifying coniferous (72.5%) and deciduous (66.3%) classes independently. In Table 3, we provide an example error matrix created using the 55% training sample size and feature reduction method, with five classes, to further understand the difference in accuracy between this approach and the accuracy achieved using the generalized (3) classes.

UAS Imagery
The UAS-SfM processing for this study generated a natural color (SODA) and multispectral (Sequoia) orthomosaic for each of the four properties. These spatial models comprised pixel sizes (i.e., ground sampling distances or spatial resolution) ranging from 11.6 cm to 13.2 cm for the multispectral imagery. The average spatial resolution was 12.55 cm. For the natural color imagery, the spatial resolution ranged from 2.53 to 3.26 cm, with an average pixel size of 3.02 cm. A number of supervised, digital, classification techniques were employed to assess forest health classes (Table 4). This table demonstrates that the highest average overall accuracy was produced using a 55% training, 45% validation, sample split and the OOB evaluation method (65.43%). This result was only slightly higher, 0.376%, than the 55% training and feature reduction method. This feature reduction was based on the MDI scores found using this method (Figure 3). We additionally applied these classification methods without the SODA green and red bands, and again without any of the SODA bands. Both iterations produced a slight decrease in the average overall accuracy. When exchanging the random forest classifier for the SVM classifier, the overall accuracy lowered by approximately 8%. Lastly, when generalizing the health assessment to 'healthy', 'stressed', and 'degraded' trees, the overall accuracy reached 71.19%. When examining one of the error matrices produced using the five-class health assessment (Table 5) it is observed that some of the misclassification was the result of confusion between coniferous and deciduous classes.

Discussion
The invasion of forest ecosystems by exotic diseases and insects is one of the most detrimental threats to their stability and productivity [108,109]. Forest health and forest degradation, known to guide losses in species diversity and timber resources, are increasingly coming to the attention of forest managers [110][111][112]. These negative effects are subject to a positive feedback loop with climate change for much of the world and are further heightening the concern of forest owners and managers as they require more intense monitoring of their forest communities [4,65]. One of the most sought-after types of information pertaining to regional stressors is the distribution and environmental factors that influence forest diseases and pests [18,22,73]. In this study, UAS imagery correctly classified forest health classes with an overall accuracy that was 14.93% higher than high-resolution airborne imagery. The lowest class specific producers' accuracy was for stressed deciduous trees. Many of these trees were incorrectly labeled as healthy. The redundancy  Table 5. Forest health thematic map accuracy assessment error matrix produced using the UAS imagery, random forest (RF) classifier, and feature reduction digital classification method. The classes represented in this error matrix include: coniferous (C), deciduous (D), coniferous stressed (CS), deciduous stressed (DS), and dead/degraded (Snag).

Discussion
The invasion of forest ecosystems by exotic diseases and insects is one of the most detrimental threats to their stability and productivity [108,109]. Forest health and forest degradation, known to guide losses in species diversity and timber resources, are increasingly coming to the attention of forest managers [110][111][112]. These negative effects are subject to a positive feedback loop with climate change for much of the world and are further heightening the concern of forest owners and managers as they require more intense monitoring of their forest communities [4,65]. One of the most sought-after types of information pertaining to regional stressors is the distribution and environmental factors that influence forest diseases and pests [18,22,73]. In this study, UAS imagery correctly classified forest health classes with an overall accuracy that was 14.93% higher than highresolution airborne imagery. The lowest class specific producers' accuracy was for stressed deciduous trees. Many of these trees were incorrectly labeled as healthy. The redundancy in the green and red image bands when using a composite of the SODA and Sequoia sensors did not have a negative influence on the classification accuracy. Instead, using all the image bands from both sensors resulted in a 1.52% increase in overall accuracy. Additionally, the MDI test conducted during the classification of the UAS imagery showed that the spectral indices (e.g., NDVI and NGRDI) were some of the most important image features along with the red edge band, which is unique to the Sequoia sensor. These results are in agreement with several other studies [53,113,114]. Lastly, when the forest health classes were generalized to 'healthy', 'stressed', and 'degraded', to avoid species misclassification, the UAS still outperformed the airborne imagery. In addition to the higher classification accuracies for forest health presented here, UAS have several additional advantages over airborne imagery. These advantages include: the ability to achieve higher species specific classifications and the ability to fly with a far greater temporal flexibility [81,115,116]. For example, the freely available airborne imagery used in this study is only collected once every three years and is not collected during consistent seasons throughout the state [80,117]. During an independent analysis of coniferous and deciduous species, the UAS imagery reached an overall classification accuracy for forest health of 71.19%. In similar studies, UAS imagery was used to assess specific tree species and disturbance types. In a study by Nasi et al. [118], a hyperspectral sensor was used to survey Norway spruce (Picea abies L. Karst.) that had been infested by European spruce bark beetles (Ips typographus L.). Their evaluation resulted in an overall accuracies of 79% for airborne imagery and 81% for UAS imagery for similar forest health classes to our study [118]. In Cardil et al. [119] researchers studied two pine dominated areas experiencing defoliation due to pine processionary moth (Thaumetopoea pityocampa). Using only a natural color camera onboard a UAS, tree level defoliation was correctly identified with an overall accuracy of 79% [119]. Time relevant, field-based surveys of forest health at actionable scales incurs too high of a cost, emphasizing the need for remote sensing tools [120]. Many contemporary investigations focus on one or two specific tree species or stressors. New England forests, however, feature a multitude of natural and anthropogenic disturbances as well as an exceptionally high species diversity at local scales [20,22,73]. A competent management tool for land managers in this region should be able to identify stressed or degraded individual trees from among the species rich population that is naturally present. An example of this spatial information is presented below in Figure 4. This map presents the individual trees assessed during the field assessment of individual tree health as well as the individual trees classified using the UAS multispectral imagery as stressed/degraded. Using this information, additional image interpretation and field surveys could be prioritized for areas exemplifying the highest severity of degradation or areas featuring the highest density of important tree species.  Despite the successes that this research and similar studies have found in the application of UAS for fine scale forest health monitoring, there are several sources of uncertainty that should be further explored. Due to the variability in response that individual trees exhibit to stress, disease, or pests, other researchers have regarded UAS as serving only as a predictor of areas requiring priority management [121]. Even using a binary classification of 'healthy' or 'degraded' trees, many environmental factors in natural ecosystems may have adversely affected our 'healthy' reference trees. While methods do exist to collect field-based spectral reflectance data, which could provide a more direct comparison to UAS remotely sensed image features, these methods elicit considerable time and resources for large study areas; especially in complex, mixed-species forests [122][123][124][125]. Another source of uncertainty in this study was the reliance on the Parrot Sequoia+ mul- Despite the successes that this research and similar studies have found in the application of UAS for fine scale forest health monitoring, there are several sources of uncertainty that should be further explored. Due to the variability in response that individual trees exhibit to stress, disease, or pests, other researchers have regarded UAS as serving only as a predictor of areas requiring priority management [121]. Even using a binary classification of 'healthy' or 'degraded' trees, many environmental factors in natural ecosystems may have adversely affected our 'healthy' reference trees. While methods do exist to collect field-based spectral reflectance data, which could provide a more direct comparison to UAS remotely sensed image features, these methods elicit considerable time and resources for large study areas; especially in complex, mixed-species forests [122][123][124][125]. Another source of uncertainty in this study was the reliance on the Parrot Sequoia+ multispectral sensor. Despite the sunshine sensor and calibration plate coupled with the Parrot Sequoia+, this sensor is subject to influences of the camera temperature, atmospheric conditions, and variability in the sunshine sensor orientation during flight [126]. Prior to use for normalizing the irradiance of the multispectral images, the sunshine sensor data should be smoothed. This pre-processing would create a more radiometrically consistent estimate of reflectance across flights and especially across study areas [44,126,127]. In the original investigations of this research, we also proposed a comparison to satellite sensors with a higher spectral resolution (e.g., Sentinel-2), as a way to more fully understand the spectral properties of these forest health classes. Early on in the classification, however, it became clear that such satellite sensors lacked the spatial resolution to sufficiently address our reference trees. Figure 5 provides representation of these data sources and the scale of the individual tree observations. tispectral sensor. Despite the sunshine sensor and calibration plate coupled with the Parrot Sequoia+, this sensor is subject to influences of the camera temperature, atmospheric conditions, and variability in the sunshine sensor orientation during flight [126]. Prior to use for normalizing the irradiance of the multispectral images, the sunshine sensor data should be smoothed. This pre-processing would create a more radiometrically consistent estimate of reflectance across flights and especially across study areas [44,126,127]. In the original investigations of this research, we also proposed a comparison to satellite sensors with a higher spectral resolution (e.g., Sentinel-2), as a way to more fully understand the spectral properties of these forest health classes. Early on in the classification, however, it became clear that such satellite sensors lacked the spatial resolution to sufficiently address our reference trees. Figure 5 provides representation of these data sources and the scale of the individual tree observations. Spatial Resolution and the Scale of Individual Tree Analysis UAS (11.6 cm pixel) NAIP (60 cm pixel) Sentinel-1 (10 m pixel) Instead of a comparison between UAS and other remote sensing platforms, data fusion remains a promising expectation for future research with these complex forests. The constraints of frequent monitoring make piloted aircraft techniques logistically challenging [58]. Image fusion allows users to overcome the shortcomings of single data source limitations. For example, with the fusion of satellite and UAS imagery, users could overcome the low spatial resolution of most satellite sensors and the limited coverage that can be accomplished by UAS [57,59]. Lastly, using UAS as an intermediate step for groundlevel observations could also increase the efficiencies found in data scaling [72,120]. UAS, as opposed to field measurements, allow for a far greater abundance of reference measurements to be made for scaling models [71]. These imagery combinations may help monitor fine scale change patterns over diverse ecosystems [128]. To accurately engage methods of data fusion between UAS and other sensors, several challenges should be examined. The first being spectral intercalibration. Despite independent radiometric calibration of the UAS data, there can remain differences between the spectral values measured by the UAS and satellite data [59]. Another fundamental challenge is the co-registration of such high-resolution imagery. Events with real-time kinematic (RTK) receivers on the misalignment of either data source by mere pixels could have a dramatic impact on the accuracy of their resulting data product [44,71,128]. Lastly, there is a consequential challenge in collecting imagery from both data sources on the same date. Even with only a few days of separation between collecting such UAS and satellite images, differences in spectral Instead of a comparison between UAS and other remote sensing platforms, data fusion remains a promising expectation for future research with these complex forests. The constraints of frequent monitoring make piloted aircraft techniques logistically challenging [58]. Image fusion allows users to overcome the shortcomings of single data source limitations. For example, with the fusion of satellite and UAS imagery, users could overcome the low spatial resolution of most satellite sensors and the limited coverage that can be accomplished by UAS [57,59]. Lastly, using UAS as an intermediate step for ground-level observations could also increase the efficiencies found in data scaling [72,120]. UAS, as opposed to field measurements, allow for a far greater abundance of reference measurements to be made for scaling models [71]. These imagery combinations may help monitor fine scale change patterns over diverse ecosystems [128]. To accurately engage methods of data fusion between UAS and other sensors, several challenges should be examined. The first being spectral intercalibration. Despite independent radiometric calibration of the UAS data, there can remain differences between the spectral values measured by the UAS and satellite data [59]. Another fundamental challenge is the co-registration of such high-resolution imagery. Events with real-time kinematic (RTK) receivers on the misalignment of either data source by mere pixels could have a dramatic impact on the accuracy of their resulting data product [44,71,128]. Lastly, there is a consequential challenge in collecting imagery from both data sources on the same date. Even with only a few days of separation between collecting such UAS and satellite images, differences in spectral reflectance, solar/viewing angles, or environmental conditions could cause inconsistencies in the data fusion process [57,128]. Another limitation of this study is the rigid and coarse temporal resolution of the airborne imagery. While attempts could be made to align the collection of the field assessments and the UAS imagery for proper comparison, the NAIP imagery is only collected and distributed every couple of years, and may not have been collected during a season in which analyzing a variety of forest disturbances is possible.
Future research could benefit from more closely matching the collection field, airborne, and UAS data, as well as the response of individual trees to forest disturbance.

Conclusions
The distribution and severity of forest health stressors present too great of an impact on natural ecosystems for field-based monitoring to capture and monitor alone. These events are causing billions of dollars in diminished ecosystem services and management costs across a variety of keystone tree species. Unmanned aerial systems (UAS) provide forest and natural resource managers with the ability to evaluate and monitor individual trees across scales that are consistent with their silvicultural practices. In our study, we examined the viability of UAS for classifying various levels of forest health within complex, mixed-species forests in New England. This assessment of UAS image analysis was completed by making the comparison of these UAS multispectral imagery to multi-factor assessments of crown vigor in the field. These results serve as a basis for prioritizing field investigations of stands identified to consist of stressed or degraded trees throughout mixed-species forests. Using a composite of natural color and multispectral UAS imagery, we achieved overall classification accuracies ranging between 65.43% and 71.19%. Some limitations in our approach include the imprecise calibration of our multispectral imagery and the variation on characteristics found among 'healthy' trees in natural environments. A necessary next step for this research is the fusion, rather than comparison, of these UAS with more widely available remotely sensed imagery. Such a step would expand the operational feasibility of UAS and address many of the challenges in precision forest health monitoring and management.   Table A1. Reference samples collected for forest health survey classes using both field methods and photo interpretation. Both coniferous and deciduous trees of the 'Healthy', 'Stressed', and 'Dead/Degraded' classes collected as reference data using both sampling methods are provided as a guide to their similarity. Appendix A Table A1. Reference samples collected for forest health survey classes using both field methods and photo interpretation. Both coniferous and deciduous trees of the 'Healthy', 'Stressed', and 'Dead/Degraded' classes collected as reference data using both sampling methods are provided as a guide to their similarity.

Conifer: Field Survey
Conifer: Photo Interpretation Deciduous: Field Survey Deciduous: Photo Interpretation Table A2. Image object features created using eCognition for the purpose of forest health classification using (1) UAS and (2) NAIP segmented imagery.

Image Classification Features
Geometric Texture Spectral