1. Introduction
Spectral radiometers (spectrometers) reach beyond the capabilities of human vision and enable scientists to retrieve diverse information from reflected light. Field spectroscopic measurements have a long history [
1] and are nowadays a common investigative tool in various research areas. Moreover, spectral vegetation analysis from air- or spaceborne platforms is a mature technology, and is commonly used for the accurate derivation of land cover classes [
2]. Hyperspectral measurements, which consist of continuous narrow spectral bands, help to retrieve information about the biophysical and biochemical components of vegetation [
3,
4,
5] and may be used to discriminate healthy or stressed plants [
6,
7].
With their synoptic view, airborne and spaceborne imaging sensors typically capture a large swath. Discrete image elements (pixels) located in the geometric center of an image are commonly acquired from a nadir view angle, whereas pixels at image edges are recorded from oblique angles. Off-nadir view geometry depends on the field of view (FOV) specifications and measurement methodology and varies among sensor systems; MODIS, for example is imaging ±55° off nadir [
8].
The bidirectional reflectance distribution function (BRDF) is the conceptual framework that explains changes in reflectance that result from view angle changes dependent on surface property and illumination [
9,
10]. BRDF influence is not desirable in a nadir image, as it impacts reflectance values recorded by the sensor and complicates the compositing of multiple images or flight lines. However, angular or off-nadir imaging can complement nadir image data by integrating additional spectral information. In forest environments, for example, an oblique view will—depending on the stand density—detect less reflectance from tree crowns and more from tree trunks [
11,
12]. Lack of knowledge in effects created from different sun-sensor geometries throughout the vegetation season have for instance led to incorrect greening estimates from satellite data in the Amazon rainforest, as recently shown by Morton
et al. [
13].
The need for BRDF correction, along with an interest in angular characteristics, has led to the development of various goniometric measurement approaches. These are able to exploit a center point from multiple view angles. The most common approach utilizes a semi-automated goniometer equipped with a point spectrometer with a radius of one meter or larger [
14,
15,
16]. On larger scales the POLDER and MISR instruments and the orbiting sensor Chris/Proba are capable of retrieving spectral data of the same area from different angles during one or multiple overpasses [
17,
18]. On a smaller scale Comar
et al. [
19] used a conoscope to assess the BRDF of wheat at leaf surface level. This technique allows characterizing the reflectance of small leaf structures, such as veins. Such multiangular measurements are necessary to accumulate knowledge regarding vegetation cover BRDF characteristics. The fundamental goal of these research efforts is to develop a model capable of predicting the BRDF of a known vegetation cover type as well as the other way round, to derive knowledge about unidentified vegetation cover from multiangular measurements. Various models have been introduced in the past to estimate BRDF on a mathematical or empirical basis [
20,
21], or to compute the aggregate energy balance of a vegetation canopy including radiative transfer, as done in the SCOPE (Soil-Canopy-Observation of Photosynthesis and the Energy balance) model [
22].
Using these methods, an effective theoretical understanding of the BDRF was developed for flat and accessible land cover like snow or soil [
23]. The small size of common goniometers along with their small FOV made the BRDF characterization of other important land cover types (including forest or agriculture) difficult [
24] or impossible. Forest and agriculture land covers are of particular significant scientific and economic interest, and alternative analytic approaches are necessary to allow BRDF measurements on larger scales and within inaccessible areas.
Some recent studies have investigated UAVs as a novel platform for goniometric measurements. Burkhart
et al. [
25] performed a survey over ice fields using a fixed-wing UAV equipped with an on-board spectrometer. Principally due to maneuvering and incident wind, the flight patterns of this platform introduced banking levels of up to ±30°, causing the spectrometer to collect multiangular hyperspectral measurements of numerous points that were overflown. A more defined method was presented by Hakala
et al. [
26] and Honkavaara
et al. [
27], who deployed a rotary-wing UAV equipped with a stabilized gimbal mounting RGB and multispectral camera, respectively. Utilizing specific flight patterns, multiangular information could be derived in the bands of the given camera.
To fully understand the BRDF effects of vegetation, we suggest that an optimized dataset providing a comprehensive understanding of multiple agricultural sites would consist of frequent multiangular hyperspectral measurements acquired at a number of different locations throughout a complete vegetation phenological cycle. Only airborne platforms can fulfill these requirements without disturbing crop growth by physically stepping through the field or casting shadows within the sensor FOV. With their recent development and improving utility and stability, UAVs can be employed as platforms for multi-angular remote data collection.
The main focus of this study is to introduce a way of collecting multiangular hyperspectral data over almost every kind of terrain and scale with a flying spectrometer. The approach combines the benefits of goniometers equipped with a high-resolution spectrometer and the flexibility of UAV platforms. We then demonstrate the acquisition and analysis of a datasets to explore BRDF effects over wheat. The angular dependency of reflectance as measured with the UAV goniometer was also compared to the reflectance modeled by SCOPE.
2. Material and Methods
The Falcon-8 octocopter UAV (Ascending Technology, Krailing, Germany) was used in this study. This platform was chosen due to its accurate flight controls and inherent stability. A hyperspectral measurement system was integrated on the UAV [
28]. This instrument was recently developed at the interdisciplinary Research Center Jülich (Forschungszentrum Jülich GmbH) and is based on the STS-VIS spectrometer (Ocean Optics Inc. Dunedin, USA). The FOV of this spectrometer is approximately 12°; spectral resolution was at a full width at half maximum (FWHM) of 3 nm, with 256 spectral bands (4 pixel spectrally binned) within the range of 338 to 823 nm.
The Falcon-8 was originally designed as a camera platform for photographers and video production. It is equipped with a camera mount whose angle can be set during flight within 1° increments. The vertical angle (tilt) is defined by the camera mount, while the horizontal angle (heading) is determined by UAV orientation. The position and navigation is done by combining the GPS information from a navigation grade GPS (Ublox LEA 6T) and the information of the orientation information of the sensors onboard the UAV. Wind gusts during the flight are counteracted by an active system, which stabilizes the camera by pitch and roll. The spectrometer is also equipped with a RGB camera, which feeds a live video stream to the operator to facilitate operation and allow proper aiming of the system.
Airborne hyperspectral target reflectance measurements were performed with the UAV spectrometer wirelessly synchronized with a second spectrometer on the ground. Latter measured a white reference (Spectralon
®) to adapt to changing illumination. A thorough calibration of the hyperspectral system was performed following the procedure described by Burkart,
et al. [
28]. This process included dark current correction [
29], spectral shift, dual spectrometer cross-calibration and additional quality checks using the SpecCal tool [
30]. Our approach allows to compute the ratio of light reflected by the target surface to the hemispherical illumination (diffuse-, ambient-, and direct-sunlight) as reflected by the white reference and is termed a hemispherical/conical reflectance factor. The actual BRDF is thus only approximated by this approach. Schaepman-Strub
et al. [
10] provide a comprehensive BRDF description and nomenclature.
2.1. Flight Pattern
Grenzdörffer and Niemeyer [
31] demonstrated that a distinct hemispherical flight pattern is necessary to enable goniometric measurements using an UAV-based airborne RGB camera. The flight pattern accurately defines the position of the UAV as well as the aiming of the camera. The flight path of the UAV is selected to follow waypoints (WP) in a hemisphere and the angle and heading of the spectrometer is set to continuously point towards the center of the hemisphere. In this manner the center of the hemisphere is measured from different viewing angles.
To quickly compute such flight patterns for UAVs, we developed the software mAngle. It was written in the platform independent open source language “Processing” and is freely available as source code and compiled versions [
32]. mAngle calculates the desired WP around a given center GPS coordinate. Placement of the WP are optimized for speed, as the UAV can quickly change horizontal position but requires more time to climb vertically to a different altitude. Flight pattern parameters including number of WP, initial angle, and hemisphere diameter can be set as desired (
Figure 1). A designated flight pattern can be exported as a *.kml file to Google Earth (Google Inc., Mountain View, CA, USA) for visualization. The flight pattern can also be exported as a *.csv file, the format used by the Falcon-8 flight planning software (AscTec Autopilot Control V1.68). Such a hemispheric flight pattern is also useful to acquire pictures around a center object of interest for 3D reconstruction.
Figure 1.
Graphical user interface of the mAngle software with input fields for the desired waypoint pattern. By setting radius, number of desired waypoints as well as starting angle and other parameters, a distinct goniometric flight pattern can be generated. A draft of the waypoint pattern is visualized in the right box of the program window.
Figure 1.
Graphical user interface of the mAngle software with input fields for the desired waypoint pattern. By setting radius, number of desired waypoints as well as starting angle and other parameters, a distinct goniometric flight pattern can be generated. A draft of the waypoint pattern is visualized in the right box of the program window.
2.2. Accuracy of the Unmanned Aerial Vehicle (UAV) Goniometer
To assess the positioning and pointing accuracy of the UAV goniometer the spectrometer was replaced by a high resolution RGB camera (NEX 5n, Sony, 16 mm lens) mounted on a similar active stabilized gimbal. In this configuration the UAV was flown following the same waypoint pattern as was used for a multi-angular spectrometer flight. In operation, the airborne spectrometer is triggered three times at each WP. The RGB camera also acquired three digital images at each WP (84 in total). Eleven ground control points (GCPs) were distributed within the covered area and registered using a differential GPS (Topcon HiPer Pro, Topcon). 3D reconstruction software (Agisoft Photoscan, version 1.0.4) was used to structure the spatial arrangement of the scene and georeference it with the GCPs. This rendering was calculated with a resolution of 3.53 mm/pixel and an average error of 1.46 pixels. The camera position and view angles for each individual image were exported and served as an estimator for the spatial accuracy of the UAV under operational conditions.
2.3. Field Campaign
Two multiangular flights (referred to as MERZ1, MERZ2) were conducted over farmland (Lat 50.93039, Lon 6.29689) on 18 June 2013 during the ESA-HyFlex campaign in Merzenhausen, Germany. The two flights were performed under cloud-free moderate wind (1.6–5.5 m/s) conditions with an interval of two hours—one hour before and one hour after solar noon (
Table 1). At the time of the study, the field contained mature wheat, with fully developed but still green ears (
Figure 2). The centroid of the hemispherical waypoint pattern was located within the field in an area of uniform cover, avoiding farm equipment tracks and trails. The center point was defined using aerial imagery, in order to avoid disturbing measurements by walking into the area of interest. The two datasets produced in this campaign are freely available via SPECCHIO [
33] at the server of the University of Zuerich under the campaign name “Merzen”.
Figure 2.
Wheat (Triticum aestivum) at the study site Merzenhausen, Germany, at the time of the multiangular flights, 18 June 2013. Ears were fully developed but still green.
Figure 2.
Wheat (Triticum aestivum) at the study site Merzenhausen, Germany, at the time of the multiangular flights, 18 June 2013. Ears were fully developed but still green.
Table 1.
Local time and duration with the corresponding sun angle parameters for the two hyperspectral flights performed over wheat field in Merzenhausen, Germany.
Table 1.
Local time and duration with the corresponding sun angle parameters for the two hyperspectral flights performed over wheat field in Merzenhausen, Germany.
Flight | Start Time | Duration | Sun Azimuth | Sun Elevation |
---|
MERZ1 | 12:43 | 09 min | 155° | 61° |
MERZ2 | 14:47 | 11 min | 213° | 59° |
For these flights a hemisphere with a radius of 16 m was specified. The spectrometer has a FOV of 12°. The areal coverage of each measurement is a function of sensor tilt angle, encompassing here 9 m
2 at nadir up to 30 m
2 with 20° tilt. WPs around the hemisphere were set to cover vertical tilt angles of 90° (nadir), 66°, 43° and 20°, at 8 equally distributed heading angles, potentially producing a total of 28 WPs. However, nadir measurements were only acquired at four different headings, which were then merged into a single WP, leading to a total of 25 WPs included in the analysis. In the following individual WPs will be identified as WP (tilt degree, heading degree). The spectrometer was activated three times at each WP to allow averaging and assessment of response variance. MERZ1 required a flight time of nine minutes, and MERZ2 required eleven minutes to consecutively measure the WP pattern. An additional UAV flight was conducted over the target using an RGB camera (NEX 5n, Sony Corporation, Minato, Japan, 16 mm lens) to image each WPs (
Figure 3).
2.4. Data Preprocessing
Each spectrum captured from the UAV was transformed to reflectance using the reference spectra simultaneously measured by the ground spectrometer. Then, for each WP, the mean, standard deviation and coefficient of variation were calculated from the three measured spectra. All further analyses were based on the mean spectra. To analyze the data with regard to the tilt and heading angle, averaged values were calculated depending on the parameter of interest. Additionally, to analyze relative changes in reflectance, spectra from all measurement positions were normalized using the nadir spectra response values [
34]. The resulting normalized nadir anisotropy factor (ANIF
band) produces a coefficient for each band, which individually adjusts (increases or decreases) reflectance factor values for each spectral band in relation to those recorded at nadir (Equation A). Thus, an ANIF factor of one describes an identical reflectance as recorded for a given band at nadir, while values above or below one describe higher or lower reflectance than the nadir value.
Figure 3.
Example Red-Green-Blue (RGB) images with tilt angles of 20°, 66° and 90°. These images were acquired at the Merzenhausen site at approximately 13:30 following a multiangular flight path identical to the spectrometer flights. The Field-Of-View (FOV) of the RGB camera is 73.7° × 53.1° (compared to the 12° FOV of the airborne spectrometer) and allows observing multiangular effects within a single image–the bright hotspot with the shadow of the unmanned aerial vehicle in the center, located in the lower left corner of the 90° image is an example.
Figure 3.
Example Red-Green-Blue (RGB) images with tilt angles of 20°, 66° and 90°. These images were acquired at the Merzenhausen site at approximately 13:30 following a multiangular flight path identical to the spectrometer flights. The Field-Of-View (FOV) of the RGB camera is 73.7° × 53.1° (compared to the 12° FOV of the airborne spectrometer) and allows observing multiangular effects within a single image–the bright hotspot with the shadow of the unmanned aerial vehicle in the center, located in the lower left corner of the 90° image is an example.
2.5. Vegetation Indices
Broadband vegetation indices (VIs) have an extensive history in remote sensing. Together with their hyperspectral counterparts they are still widely used in vegetation studies [
35,
36]. VIs commonly ratio near-infrared (NIR) and red band reflectance values in order to compensate for influences of different illumination conditions or background materials. To investigate the effect of the BRDF we examined three common Vis (
Table 2) and calculated their values for all WPs. The Normalized Difference Vegetation Index (NDVI) uses two wavelengths in the red and NIR domain and has been widely used in a diverse range of applications. In our study we used the NDVI as proposed by Blackburn
et al. [
37]. As a second index we used the Transformed Chlorophyll Absorption in Reflectance Index (TCARI) developed by Haboudane
et al. [
38]. TCARI was developed to predict chlorophyll absorption and uses wavelengths in the green, red and NIR spectral regions. The last index used in this study is the Red Edge Inflection Point (REIP). Originally introduced by Guyot
et al. [
39] it characterizes the inflection in the spectral red edge by calculating the wavelength with maximum slope. It has been used to quantify leaf chlorophyll content [
40].
Table 2.
Vegetation indices used in this study and their underlying formulas.
Table 2.
Vegetation indices used in this study and their underlying formulas.
Index | Formula | Reference |
---|
NDVI | (R800 − R680)/(R800 + R680) | Blackburn et al. 1998 |
TCARI | 3 × ((R700 – R760) – 0.2 × (R700 – R550) × (R700/R670)) | Haboudane et al. 2002 |
REIP | 700 + 40 × (((R667 + R782)/2) – R702)/(R738 – R702)) | Guyot et al. 1988 |
2.6. Data Visualization
Several different visualizations or graphics were used in this study to focus on specific features under investigation. An effective method for assessing multiangular measurements includes the use of a segmented circular display known as a “polar plot”. The polar plot shown here in
Figure 4 represents the UAV headings and sensor tilt angles within a circular matrix and illustrates the intensity of the measurement values by applying a color to each segment. To provide an useful overview of the dataset of this study and to include as well a comparison of the reflectance in the spectral domain, multiple plots are necessary.
Figure 4.
Reflectance of wheat at 480 nm measured at all 25 waypoints shown as a circular graph, or polar plot. Each “slice” represents a heading while each ring represents a sensor tilt angle. Spectral reflectance magnitude is color coded from low values of light blue, to high values in bright red. The angular position of the sun is depicted by the sun-symbol. In this figure no interpolation between waypoints is performed.
Figure 4.
Reflectance of wheat at 480 nm measured at all 25 waypoints shown as a circular graph, or polar plot. Each “slice” represents a heading while each ring represents a sensor tilt angle. Spectral reflectance magnitude is color coded from low values of light blue, to high values in bright red. The angular position of the sun is depicted by the sun-symbol. In this figure no interpolation between waypoints is performed.
2.7. Radiative Transfer Model Comparison
To compare the multiangular UAV measurements to modeled data, the SCOPE radiative transfer model was tested. The model generates the spectrum of outgoing radiation in the viewing direction as a function of vegetation structure [
22]. SCOPE input parameters were derived through comparison of the MERZ1 nadir spectrum with a lookup table of SCOPE spectra generated using a permutation of input parameters that were expected from wheat at the present phenological state. The resulting best-fit parameters are shown in
Table 3.
Table 3.
Soil-Canopy-Observation of Photosynthesis and the Energy balance model (SCOPE) input parameters: Leaf Area Index (LAI), Leaf Inclination (LIDFa), Chorophyll A/B (Cab) content in µg/cm2, Leaf Thickness Parameter (N), Leaf water equivalent layer (Cw) in cm, Dry matter content (Cdm) in g/cm2, Senescent material fraction (Cs), Variation in leaf inclination (LIDFb). Default values were used for all other SCOPE input parameters.
Table 3.
Soil-Canopy-Observation of Photosynthesis and the Energy balance model (SCOPE) input parameters: Leaf Area Index (LAI), Leaf Inclination (LIDFa), Chorophyll A/B (Cab) content in µg/cm2, Leaf Thickness Parameter (N), Leaf water equivalent layer (Cw) in cm, Dry matter content (Cdm) in g/cm2, Senescent material fraction (Cs), Variation in leaf inclination (LIDFb). Default values were used for all other SCOPE input parameters.
Fitted Parameters | Constant Parameters |
---|
LAI | LIDFa | Cab | N | Cw | Cdm | Cs | LIDFb |
3.5 | −0.35 | 95 | 1.5 | 0.004 | 0.005 | 0.15 | −0.15 |
Using the input parameters above, the angular module of SCOPE was run to estimate the reflectance spectra at identical angles as those measured with the UAV goniometer. Sun azimuth and zenith angles were set to match the values present at the time of the MERZ1 measurements.
4. Discussion
In this study, a new goniometer approach for large-scale measurement of BRDF is presented and an initial hyperspectral dataset is analyzed. By deploying the spectrometer on a rotary-wing UAV there is no longer a need to mount the instrument on large ground-based positioning structures. The large FOV has the advantage of averaging out small variations, which are part of the canopies variability. As the device is flying, surfaces can be investigated with desired measurement patterns even over areas inaccessible by land and without disturbing the eventual surface cover like vegetation. Until now these areas had to be approached using satellites or by modeling [
41]. In remote sensing applications, where goniometers cannot be deployed, angular effects are currently minimized or correction approaches are applied: Field-spectrometer measurements are carried out around the same time (noon) and from nadir view [
5]. Thus the sun-object-sensor geometry is almost stable. For UAV-, air- and spaceborne systems a number of correction methods have been developed. These include the use of image statistic based methods for flat terrain [
42] and physical or semi-empirical models such as for the processing of MODIS data [
43]. Lately, a more generic BRDF correction method was introduced, which builds on a surface cover characterization [
44]. Since physical and empirical models are based on the current knowledge and BRDF effects depend on many factors, the flying goniometer could help to evaluate and eventually improve the correction methods.
We assessed the pointing accuracy of the UAV system and found it to be of acceptable accuracy for a GPS-aided flying system, although it is still not as precise as ground-based instrumentation [
34,
45]. Parameters of altitude, X/Y position and sensor tilt angle are highly accurate within the navigation grade GPS specifications. Additionally, the remounting of the RGB camera described in
Section 2.2 might have introduced an artificial error. The relative position as described by the low standard deviation demonstrates the precision of the system (
Table 4). However, the vehicle heading parameters are less accurate. Relative heading inaccuracies may be ascribed to the Falcon-8 flight control system, which does not make use of a magnetic compass. If operated in an environment with a strong magnetic field, a compass system could produce serious errors in vehicle position and heading readings and cause catastrophic UAV failures. However, in the case of the UAV goniometer, the accuracy would improve through the use of a compass system for heading correction.
Additional sources of error in the platform/sensor system are found in gimbal calibration in the tilt axis and during the process of physically mounting the spectrometer on the gimbal. Inconsistencies in either one or both of these procedures will lead to pointing offsets. The system could also be improved by deploying the spectrometer and a RGB camera in tandem, triggering both simultaneously. Camera and spectrometer could be aligned and calibrated in the laboratory to determine the spectrometer field of view in the camera image. The camera imagery could then be utilized to accurately calculate the position and pointing of the UAV using the structure from motion approach used in this study to evaluate the pointing accuracy (
Section 2.2).
The UAV system in combination with the “mAngle” software enables users to plan, setup and perform a multiangular flight around a center point of interest efficiently and quickly (in less than 30 min, 10 min for the measurement flight itself). In addition, the UAV and spectrometer system is deployed in a single, easily portable package, making it highly mobile. Since the completion of this study, the system has been deployed at a number of other sites in Europe and New Zealand.
The large radius and thus big footprint of the UAV ensures a good averaging over the fine structure of the vegetation (e.g., leaves, shaded areas, stems). This was assessed by calculating standard deviation of multiple spectra at the same waypoint (
Figure 6) and shows good agreement of consecutive spectra taken at the same WP. If smaller footprints are desired the flying radius can be reduced or the spectrometer can be equipped with a fore optic with a narrower FOV.
The results of the multiangular reflectance measurements acquired in this study are consistent with previous measurements characterizing common angular reflectance distribution over vegetation [
46]. The common hotspot feature is clearly visible in the data and changes over time with sun angle. High levels of reflectance were found at the rather low tilt angle of 20° in the heading of the hotspot. As the tilt angle is lower than the hotspot feature, these high levels might be introduced by a viewing angle, whereas only the very top of the canopy is seen by the sensor (
Figure 9). Along with the results of our accuracy assessment of UAV imagery pointing and of the spectral domain response, we are confident that we have utilized a novel platform-sensor combination to acquire a valid and valuable hyperspectral dataset.
The complete spectrum analysis emphasizes that BRDF effects are both wavelength and angle dependent. Around the hot spot the measured reflection is higher than in nadir, in both in the VIS and NIR part of the spectrum. For WPs towards the dark spot the reflectance is lower in most parts of the VIS and higher in the NIR. Overall, lower sensor tilt angles increase reflectance compared with the nadir position. While the NDVI reduces angular effects quite efficiently, these effects were a significant factor in TCARI. This distinction can be ascribed primarily to the differing formula structures of the two VIs. For the NDVI, the reflectance in the NIR dominates the nominator of the formula. Thus the differences due to the observation angle influence the index nominator and denominator in similar ways and the entire ratio only slightly changes. The TCARI formula does not provide such normalization. The reflection factors at the wavelength of the first part of the term (R700–R760) are differently influenced by the angle (
Figure 8) and introduce strong fluctuations to the VI. Minor influences are introduced by the second term. The first part (R700–R550) of the second term is not strongly influenced, since both reflection factors of the wavelengths are affected similarly by the angle. However, the second component of TCARI again uses the R700 and R760 band ratio. This increases the variations in the second term of the formula caused by the differing observation angles. In combination, these factors produce the significant differences (up to 150 percent), which are seen in the TCARI values. Differing observation angles cause only minor fluctuations in REIP values. As seen with NDVI, formula deviation normalizes most of the variation in REIP values. However, it must be emphasized that, as for most VIs, the practical dynamic range of the REIP is narrower than what is theoretically possible. Thus even the small observation angle variances suggested by the REIPs results could lead to errors in interpreting this index if BRDF effects are disregarded. Other studies have been carried out for other VIs or vegetation cover [
47,
48,
49,
50] support the angular dependency found in this study.
Radiative transfer models show significant potential as tools for correcting angular influences introduced by solar effects or imaging sensors. They are based on existing theory of radiative transfer and plant physiology [
22]. So far, real world multiangular data for various vegetation covers are rare and thus, a rigorous validation of the model is challenging. With the approach described in this study datasets for the validation and improvement of those methods may be generated. However, it has to be taken into account that SCOPE does not account for certain sensor variables such as FOV and FWHM. Due to the footprint of the spectrometer, light reflected at different angles by the canopy is collected by the sensor. Thus, this might be one source for the increasing discrepancies towards low tilt angles observed in this study. Following studies could minimize this effect by using spectrometer fore optics with narrower FOV. A careful investigation on the difference between modeled and measured spectra go beyond the scope of this study but should be investigated in the future.
Based on this study, we strongly encourage the extensive compilation of multiangular datasets for various vegetation cover types and environments. A more sophisticated knowledge base regarding vegetation angular effects could also enable researchers to derive accurate complementary information through the use of angular measurements that capture vegetation features not typically visible from a nadir perspective [
12]. Additionally, these results could help to understand influences of BRDF effects in imaging spectroscopy. Typically, current hyperspectral (image-frame and push broom) imaging systems as well as RGB systems have a FOV of up to 50° [
35]. Thus, pixels captured towards the edges of the image have tilt angles of about 66°. As shown here, angular effects have a significant contribution to these observation angles and need to be taken into account during analysis. To improve the correction of these effects consecutive studies should examine tilt angles found in the FOV of common UAV and airborne sensors. This is foreseen within a number of parallel research activities that are ongoing and focused on improving models and collecting spectral databases. These include COST Action ES0903 EUROSPEC, COST Action ES1309 OPTIMISE, and the SPECCHIO online spectral database [
33]. These projects could also serve as a basis for enhanced training of models leading to highly accurate correction methods.