Next Article in Journal
In-Season Diagnosis of Winter Wheat Nitrogen Status in Smallholder Farmer Fields Across a Village Using Unmanned Aerial Vehicle-Based Remote Sensing
Next Article in Special Issue
Mapping an Agricultural Field Experiment by Electromagnetic Induction and Ground Penetrating Radar to Improve Soil Water Content Estimation
Previous Article in Journal
Seed Yield and Lodging Assessment in Red Fescue (Festuca rubra L.) Sprayed with Trinexapac-Ethyl
Previous Article in Special Issue
Three-Dimensional Point Cloud Reconstruction and Morphology Measurement Method for Greenhouse Plants Based on the Kinect Sensor Self-Calibration
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Unmanned Aircraft System (UAS) Technology and Applications in Agriculture

by
Samuel C. Hassler
and
Fulya Baysal-Gurel
*
Department of Agricultural and Environmental Sciences, Otis L. Floyd Nursery Research Center, Tennessee State University, McMinnville, TN 37110, USA
*
Author to whom correspondence should be addressed.
Agronomy 2019, 9(10), 618; https://doi.org/10.3390/agronomy9100618
Submission received: 6 September 2019 / Revised: 4 October 2019 / Accepted: 7 October 2019 / Published: 9 October 2019
(This article belongs to the Special Issue Precision Agriculture for Sustainability)

Abstract

:
Numerous sensors have been developed over time for precision agriculture; though, only recently have these sensors been incorporated into the new realm of unmanned aircraft systems (UAS). This UAS technology has allowed for a more integrated and optimized approach to various farming tasks such as field mapping, plant stress detection, biomass estimation, weed management, inventory counting, and chemical spraying, among others. These systems can be highly specialized depending on the particular goals of the researcher or farmer, yet many aspects of UAS are similar. All systems require an underlying platform—or unmanned aerial vehicle (UAV)—and one or more peripherals and sensing equipment such as imaging devices (RGB, multispectral, hyperspectral, near infra-red, RGB depth), gripping tools, or spraying equipment. Along with these wide-ranging peripherals and sensing equipment comes a great deal of data processing. Common tools to aid in this processing include vegetation indices, point clouds, machine learning models, and statistical methods. With any emerging technology, there are also a few considerations that need to be analyzed like legal constraints, economic trade-offs, and ease of use. This review then concludes with a discussion on the pros and cons of this technology, along with a brief outlook into future areas of research regarding UAS technology in agriculture.

1. Introduction

A wide variety of sensors and other data-gathering equipment have been developed for agricultural purposes, such as underground sensors monitoring soil quality, aboveground sensors monitoring temperature and humidity, yield sensors, and weed sensors; yet, in the realm of precision agriculture, imaging sensors are perhaps the most important [1,2,3]. Traditionally, aerial images could only be gathered from satellites, such as those from the Landsat program or through aircraft [4,5]. Using multispectral and hyperspectral cameras equipped on a satellite or aircraft, images could be used to calculate various vegetation indices which can indicate variabilities in the field. Vegetation indices, such as the normalized difference vegetation index (NDVI), rely on comparing light intensities reflected from canopies in the visual and near infra-red (NIR) range [6,7]. Given the vast distance from a satellite to a field and the area an image can cover, resulting images often have high temporal resolution and low spatial resolution compared to other methods [8]. In order to produce more detailed images with high spatial resolution at a low cost, cameras mounted upon unmanned aerial vehicles (UAVs) were utilized with promising results [9]. As these UAVs began to incorporate more peripheral technologies and grew in complexity, a new term was developed to describe the whole system together—unmanned aircraft system (UAS) [10,11]. Along with this new technology, came many new challenges such as processing of geospatial data [12,13,14] and lower temporal resolutions when applied to large areas of land [7,15]. Therefore, it should be noted that UAS technology is not meant to replace satellite data, as there are trade-offs for using one over the other. In many situations, both UAS and satellite imagery are used in conjunction [8,16].
It should come as no surprise that this review has ambitious goals, as UAS technology is becoming increasingly widespread in the modern era. Thus, this article should serve as a comprehensive introduction to the topic as it is related to agriculture, rather than a complete summation of all the research in this area. This technology, at least in the realm of agriculture, is still in its infancy, with many possible uses yet to be explored. Therefore, some concepts presented have only been partially developed for drone usage or are still in the research phase of development and are not yet available for consumer use. The primary goals of this review are to cover the popular areas of use; physical components of these systems; data gathering and processing; a few considerations such as legal, economic, and integration factors; and to summarize with a brief discussion of the advantages, disadvantages, and future areas of research in agriculture.

2. Areas of Use

2.1. Field Mapping

As field mapping is often the first step in many other areas of use, this should prove the most appropriate place to begin. Field mapping has many uses from small farm improvement to large-scale governmental land surveys [4]. A few prominent studies involving field mapping in agriculture include: NDVI maps [7,17,18,19], assessing spatial variability and structure of vineyards [20], surveying and development of land cover maps [4,15], and estimation of tree locations and sizes [21,22].
These maps are often constructed using a series of independent images taken from an UAS and stitched together using various photogrammetric software packages [15,23]. Each individual image in the resulting mosaic are commonly georeferenced using ground control points (GCPs), overlaying Landsat satellite data, or a combination of the two [4,7,15,17,18,19,20,21,22]. Images will also undergo a process of data correction which can involve vignetting and orthorectification [18], among others, depending on the task.

2.2. Plant Stress Detection

Detection of various plant stresses is a widely used application of UAS in agriculture. The most common forms of stress being detected via UAS include water stress, nutrient stress, and plant diseases. Water stress can be fairly simple to detect and is generally detected through a combination of NDVI and crop water stress index (CWSI) values, the latter of which is generated using thermal imaging [24,25]. Nutrient stress can vary in detection difficulty, as there are many more causal factors associated. As for the detection of nitrogen deficiency, there have been a few successful studies. Zaman-Allah and colleagues showed how NDVI data from a multi-spectral camera aboard a UAS can be used to phenotype low-nitrogen stress tolerance in maize [26]. Other more specific vegetation indices have been developed for more accurate detection of nitrogen deficiency remotely, such as the nitrogen nutrition index (NNI), which can be calculated using the modified chlorophyll absorption ratio index (MCARI) and modified triangular vegetation index 2 (MTVI2) obtained through hyperspectral imaging [27]. Another index used in detection of nitrogen deficiency is the Nitrogen Balance Index (NBI), which calculates the ratio of chlorophyll to polyphenols. In a paper by Li and colleagues studying rice fields in China, NBI was predicted using standard RGB imaging converted to hue, saturation, and brightness which generates a dark green color index, and which proved representative of field nitrogen concentrations [28].
Automatic plant disease detection via UAS is a field that has emerged over the past couple decades with great future research potential, as individual diseases can have drastically different symptoms across many different host species. In one study looking at Huanglongbing (HLB) detection in citrus crops by Garcia-Ruiz and colleagues, it was found that the NIR-R index value and the average reflectance values at spectral bands of 560 nm and 710 nm were statistically different between healthy and HLB-infected trees [29]. Likewise, Flavescence dorée disease in vineyards was detected using a multi-spectral camera monitoring bands at approximately 530–600–650–690–730–750–800 nm, with an accuracy over 94% by Al-Saddik and colleagues [30]. In other studies, vegetation indices were more heavily used. For example, Lowe and colleagues showed how the presence of laurel wilt disease in avocado was found to be strongly correlated with excess red (ExR) and blue/green (B/G) vegetation indices [31]. However, for some UAS-based disease detection programs, a standard RGB camera is all that is needed for accurate detection. From a study looking at phymatotrichopsis root rot (Texas root rot) in alfalfa by Mattupalli and colleagues, it was shown how RGB images alone can provide enough data for accurate classification of the disease; with the extrapolation of RGB to HSV (hue, saturation, and value) and Hrot60SV (hue values rotated 60 units, saturation, value) having almost no effect on classification [32]. Yet, in the majority of cases, RGB imaging alone is not sufficient. Zhang and his colleagues showed how the detection of sheath blight in rice could be initially classified with high-resolution RGB cameras, but multi-spectral imaging gathering NDVI data was required in order to assess disease severity [33]. It should be noted, though, that not all research in this area targets purely automated disease detection. In a recent study by Kalischuk and her colleagues, it was found that UAV-assisted scouting could provide a 20% increase in early detection of gummy stem blight in watermelons than from traditional scouting alone [34].
Although many approaches have been developed for the detection of a particular disease affecting a particular plant with great success, much of this science has yet to be applied to large-scale monitoring across many crop lines. The shear amount of data involved in this detection is pushing many researchers to look into machine learning techniques for integrated disease management through early plant disease detection. Some of these techniques including classification trees, support vector machines, and deep learning which will be discussed further in a later section [35,36,37].

2.3. Biomass and Field Nutrient Estimation

There are many areas in which precision agriculture via UAS can be applied; however, biomass and nutrient estimation is certainly one of the most popular uses for this technology. There have been numerous studies using UAS which aim to predict biomass of crops and assess nutrient levels within soil and plants alike. In a few studies, biomass yields of grasses were estimated using point clouds and crop height models (CHMs) [14,38], though, the majority of studies primarily used some form of vegetation index to make predictions [6,7,23,39,40]. Geipel and his colleagues used vegetation indices to predict both biomass and nitrogen content within plants [41]. Tokekar and his colleagues introduced a new approach to detect nitrogen levels in the soil using vegetation indices, soil sampling, and other methods through a combination of both UAS and unmanned ground vehicles (UGVs) [2]. In 2016, Aldana-Jague and his colleagues were able to map soil organic carbon levels in fields using support vector machines with data provided by a multispectral camera aboard an UAS [42]. A few common indices used to estimate biomass and nutrient levels include soil plant analysis development, NDVI, normalized difference red edge index, and optimized soil-adjusted vegetation index (OSAVI) [7,43]. Aside from vegetation indices, the use of radiative transfer models such as PROSAIL has also shown great success in estimating biomass [44]. This method relies on estimating the biophysical variable called leaf area index (LAI) or green area index (GAI) which is generally performed through the use of lookup tables; though, iterative optimization techniques and artificial neural networks are also used [45,46].

2.4. Weed Management

Many weed scientists have seen the potential of UAS and are currently working on projects to build weed maps for various agricultural crops, such as wheat and cranberry [47,48]. In 2013, a research article was published that assessed the optimal flight conditions (e.g., altitude and quantity of GCPs) needed for weed detection in young wheat fields. Their research found that an altitude of 30–100 m and roughly one GCP per image would provide the best setting for weed detection [49]. Many following studies were able to build upon this work and begin to generate weed maps for their own crops of interest. For example, Mink and his colleagues equipped a drone with RGB and multispectral cameras to measure vegetation indices and plant height data in order to detect weeds in maize and sugar beet fields. They were able to develop a weed height model which they used to detect weed instances, based upon the vegetation index excess green red (ExGR) and CHM [50]. Shortly after, Pflanz and his colleagues published an article which demonstrated how they were able to use an image classifier called Bag of Visual Words along with a support vector machine (SVM) to map weeds within a field of wheat. Flying from a much lower altitude (1 to 6 m) than previously suggested and working with a dataset of over 25,000 sub-images, they were able to detect and map instances of weeds and label them by species [51]. In 2017, Bah and his colleagues were able to apply the Hough transform technique, along with the spatial relationship of pixels, to detect weeds in row crop fields with promising results [52]. In 2018, Rasmussen and his colleagues developed a procedure for weed detection called Thistle Tool which can be implemented with only RGB imaging taken from a UAS. This method involves a process of dividing the orthomosaic into subsections, calculation of excess green index values, a classifier step, and finally the resulting weed map with segmentation of weeds and crop [53].

2.5. Counting

Though not the largest area of use, many areas of agriculture require quick and efficient counting methods for inventory purposes [54]. This method of UAS-based counting proves especially useful in citrus tree production, where the majority of related research is being conducted. In 2014, She and her colleagues developed two different algorithms to be used for counting-related work with UAS imaging. The first, which applies to crops with uniform canopies like citrus trees, used vegetation index calculations to separate tree canopies from the background and then implemented a counting method based upon the average canopy area. The second algorithm, which applied to conical-tipped Christmas tree counting, involved locating and counting local reflectance maximums within particular windows which represent the cones of trees [55]. In 2018, the same group of researchers published another paper addressing issues regarding counting difficulty in densely packed areas of plants (e.g., for shipping purposes) and provided an updated methodology using SVMs for accurate counting in these settings [56].
From a research team at IBM (International Business Machines Corporation, Armonk, NY, USA), comes another study regarding citrus tree counting using machine learning techniques. They address the problem of accurate counting in high-density fields by incorporating convolutional neural networks, a form of machine learning based in deep learning. They were able to obtain accuracy scores (upwards of 94%) similar to She and her colleagues [54,55]. Aside from plant counting, there has been some effort in the realm of fruit counting. It has been shown by Rahnemoonfar and her colleagues that computer vision algorithms are able to count ripened tomatoes with a 91% accuracy from manually taken images of tomato plants [57]. This research was furthered in 2019 by applying these methods to UAS, as well as many other realms of agriculture including live animal counting [58].

2.6. Chemical Spraying

Modern agriculture has nearly become synonymous with chemical usage, as it is crucial to almost every sector and crop [59,60,61]. Therefore, the many disadvantages involved with chemical use should be constrained as much as possible. Unmanned aircraft systems equipped for chemical spraying have been proven to eliminate many of these disadvantages such as terrain limitations of ground-based sprayers [61]; chemical exposure risks for farmers/workers [59]; and with proper use, UAS sprayers can be more effective and economical than traditional methods [59,60,62]. For example, Mink and his colleagues were able to save 90% of herbicide in maize fields and 43% of herbicide in sugar beet fields by using site-specific weed control methods via UAS [50]. From the literature, it appears that Asia is leading the way in this area of research, with a great deal of research focus upon droplet distribution and drift [60,61,62,63,64]. According to Pederi and Cheporniuk, approximately half of all Japanese rice fields currently use UAS for spraying purposes [59]. Interestingly, this is the only prominent sector of drone usage in agriculture which is not solely reliant upon cameras and imaging, though many studies find imaging tools useful for site-specific spraying [65]. It should also be noted that depending on the country there might be legal issues surrounding the use of UAS for pesticide spraying, which will be covered in the Legal subsection (Section 5.1) of this review [63].

2.7. Miscellaneous

This subsection will primarily deal with novel, interesting uses of UAS in agriculture, which do not fall into neat categories, but is included to show some of the creative possibilities with this emerging technology.
Traditionally, phenotyping for breeding trials has been conducted through time-intensive and laborious means [66]; therefore, Chapman and his colleagues conducted a study using an UAS to help alleviate this time-consuming and costly procedure in 2014. They were able to autonomously detect variations within fields of crops using RGB, infrared, and thermal cameras mounted to a UAS, which can reduce the time involved in phenotyping new lines in plant breeding operations [67]. In a similar study, tomato spotted wilt disease resistance phenotyping was conducted through an UAS using multispectral imaging and vegetation indices, which improved detection time to as early as 93 days after germination [68]. Tripicchio and his colleagues set out to improve large-scale analysis of cultivation techniques in which they successfully implemented a method to remotely detect field plowing depths without the use of satellite imaging [69]. In an unusual cross-over between military security and agricultural research, from a research group in Thailand, UAS were fitted with custom-made gas sensors and used to detect volatile compounds such as those released from cattle or ethylene from ripened fruit [70]. In 2016, Rangel developed a UAS for carrying and dropping biological control agents on field environments [71]. Finally, from a study conducted in Germany, radiation use efficiency in maize fields was calculated remotely by a UAS equipped with low-cost imagery providing insight into photosynthetic activity of crops on a per field basis [72].

3. Platforms and Peripherals

This section describes the individual physical components that come together to form entire UAS currently used in agriculture. It will cover the underlying platforms (UAV), a few commonly used peripherals (cameras, sprayers, grippers, geospatial sensors, and sense-and-avoid sensors), and computing systems to tie everything together and process data.

3.1. Platforms

The term platform, in relation to UAS, refers to the underlying UAV structure onto which the extra tools and sensors are mounted. There are numerous types of available platforms to choose from including parachutes, blimps, gliders, rotorcrafts, and fixed-wing aircrafts [63,73]. However, the most commonly used platforms in agriculture are fixed-wing aircrafts and rotorcrafts. Fixed-wing aircrafts are generally able to fly faster and cover more distance than other platforms; though, they are not suited for tasks that require a great deal of maneuvering or hovering. Due to the fact of their ability to cover vast areas of land quickly and create datasets with relatively high temporal resolution, they can be primarily found in applications such as large-scale field mapping [15,74] and occasionally, spraying [59]. Most other applications are dominated by the rotorcrafts for their ease of use (no runway is required), lower cost, and high spatial resolution imaging due to the ability to hover as seen in Figure 1 [75]. Rotorcrafts come in many forms with any number of propellers, with the most common configurations including traditional helicopters [67]; quadcopters (4 propellers) [19,68]; hexacopters (6 propellers) [21,76]; and octocopters (8 propellers) [8,72]. When deciding which platform to build upon, it is important to check for max payload capability as well as any weather restrictions—such as whether or not the platform is waterproof—in order to fit the needs of the project.

3.2. Imaging

As aerial imaging is the most common use of UAS in agriculture, it follows that there have been many imaging sensors and cameras specially developed to operate on board these systems. Every UAV platform will have a particular payload which limits the sizes of imaging equipment that can be used. Likewise, as payload increases, the UAS will suffer from decreased speed, stability, and flight time [75]. Another major consideration when equipping a UAV is the velocity at which it will be flying when capturing images. Higher flight speeds can cause motion blur in images without the proper cameras and supporting algorithms; however, this becomes less of an issue in UAS with greater hovering capabilities such as rotorcrafts [15,67,77]. A few of the important sensors and cameras that will be covered in this subsection include: standard RGB cameras, multispectral and NIR cameras, hyperspectral cameras, thermal sensors, and depth sensors [73].

3.2.1. RGB Cameras

Red, green, and blue imaging is by far the most standard form of photography available to the average person. The concept is fairly intuitive; RGB cameras produce images with three values for each pixel—red, green, and blue—measuring the intensity of the respective color. This form of imagery does not provide the same spectral detail as other options; however, a few studies have shown how standard RGB imagery can be an effective tool for 3D modeling of agricultural crops [14,21,78] and biomass estimation [6,23,38]. For 3D modeling of crops, RGB images are used to generate point clouds (see Section 4.2) of canopies and trees using methods such as object-based image analysis [78] or structure from motion (SfM) [38] with relatively high accuracy. Another modeling method using RGB imaging and point clouds relies on finding local maxima in order to locate individual trees and associated heights [21]. The other main use of RGB imagery is for biomass estimation, where NIR and multispectral cameras are often used in conjunction with RGB cameras in order to improve accuracy [6,23]. Though in a recent paper, Grüner and her colleagues were able to accurately estimate biomass in grasslands from RGB imaging alone using SfM processing [38]. Finally, in a unique approach from Mattupalli and his colleagues using a maximum likelihood classification algorithm, phymatotrichopsis root rot disease was accurately detected in alfalfa fields using only RGB imaging [32].

3.2.2. Multispectral and NIR Cameras

The most common use of multispectral and NIR cameras in agriculture is for the creation of vegetation indices which relies on NIR data or specific bands of light [19,41,68,79,80]. Near infra-red cameras work similarly to RGB cameras with internal modifications or lens filters added to tweak the bands of light captured from the traditional RGB to incorporate other bands, usually from the NIR region [72,81]. On the other hand, multispectral cameras are specifically designed to capture multiple bands (5 to 12 are common) of light for every generated pixel, instead of just three for RGB images [80]. This can be useful as plants reflect light in a wider range of light than humans can detect with the naked eye. The majority of UAS-assisted aerial imaging for detection or estimation purposes uses multispectral imaging heavily to calculate vegetation indices such as NDVI and other variants involving NIR data [6,26,33,34,41,68]. A notable exception comes from an article by Al-Saddik and her colleagues which directly used multispectral data programmed to specific bands to detect Flavescence dorée disease in vineyards without using common indices [30].

3.2.3. Hyperspectral Cameras

Similar to multispectral imaging, hyperspectral imaging is able to collect light across a wider range than traditional methods; however, the difference is in the size of the bands—hyperspectral cameras can potentially monitor thousands of narrow bands of light for every pixel in the resulting images [80,82]. This method of imaging can be useful if very specific wavelengths of light need to be measured independently—often detecting light emitted from particular biomolecules such as chlorophyll [16,27]; xanthophyll [83]; mesophyll [82]; and carotenoids [16]. The downside of hyperspectral imaging is the generally high cost of cameras (especially lightweight versions to be mounted aboard a UAV) [66,80] and the generation of overwhelming amounts of useless data if not properly calibrated by experienced professionals [40,82,84,85,86]. Therefore, multispectral cameras should be the default imaging choice unless there is a special need for the detail found in hyperspectral images.

3.2.4. Thermal Cameras

Thermal sensors and cameras are able to capture infrared light radiation in a range of approximately 0.75 to 1000 μm (for reference, NIR ends at 2.5 μm), which can provide localized temperatures of objects within a thermal image [87]. While many NIR cameras can be cheaply modified from standard RGB cameras [81], thermal cameras are specially designed to cover most or all of the infrared region [87]. Thermal imaging is mainly used in agriculture to detect stomatal regulation and, thus, water stress in crops [24,25,66,87,88]. Unfortunately, thermal imaging tends to suffer from low resolution due to the environmental variables in an open system [87], especially in the case of small uncooled thermal cameras which are the dominant form used in UAS [89]. Thermal imaging is often used in conjunction with other imaging tools (such as RGB cameras) as a form of supplementary data [24,88].

3.2.5. Depth Sensors

A final, common class of peripherals aboard UAS in agricultural-related applications are depth sensors. The RGB-D sensors provide a relatively cheap and effective way of generating depth data by capturing an extra value in each RGB pixel—a value indicating distance from the sensor to that point in the image [90]. It is true that standard RGB imaging can predict depth values; however, Wang and Li demonstrated how RGB-D cameras are able to provide greater accuracy in volume estimation than standard RGB imaging in 2014 [91]. Another popular type of depth sensor uses LiDAR technology. In contrast to RGB-D sensors which primarily rely on light reflection intensities [90], LiDAR sensors use laser pulses to map distances [75]. Common reasons to incorporate depth sensors into a UAS for imaging purposes include: altitude monitoring during spraying [92]; phenotyping [90]; and 3D modeling [93,94].

3.3. Spraying Equipment

With the continual development of UAV and UAS technology, remote pesticide or fertilizer application through these tools has been a focus of much research. Compared to crop dusting performed by aircraft, UAS have the option of applying chemicals much closer to the field which results in less potential drift [95]. However, this comes with an increased payload (chemicals, tanks, pumps, nozzles) and, thus, shorter flight times [59,60]. To address this issue, Stark and his colleagues proposed a network of small UAS to autonomously spray chemicals to reduce temporal variance [95]. Other researchers have focused more on developing methods for site-specific spraying to reduce payload upon the UAS, all of which rely heavily upon image sensing technology and machine vision algorithms [47,50,60]. Regardless, the equipment needed for chemical spraying via UAS is fairly straight forward and standardized. The UAS platform needs to be equipped with three main components: a tank to hold the liquids, a nozzle to guide the spray pattern, and a pump to move pressurized fluid through the nozzle. Common tank sizes used aboard UAVs range from a couple liters to upwards of 10 L for larger rotorcrafts and fixed-wing platforms [59,62,76], with a few setups incorporating multiple smaller tanks capable of handling different chemicals [50]. There appears to be two dominant types of nozzles used: centrifugal for a more circular pattern and flat fan for dispersal in a straight line [65]. The type of pump should be decided depending on the intensity of spraying, the amount of power available aboard the UAS, and payload factors [65,76].

3.4. Gripping Tools

Robotic arms and grippers can be seen in many applications across agriculture; however, the technology currently comes with many limitations and has yet to be implemented on a wide scale due to the lack of infrastructure and supportive software available [96]. Though still largely in the research stage, many robotic tools have been designed for usage aboard UGVs. In 2016, Gealy and colleagues provided an example with the development of a robotic device enabling the automated tuning of emitters for irrigation purposes. In this project, they built a prototype for handheld use and mentioned that the design can also be applied to a UGV [97]. Other notable areas of application are in areas such as soil preparation and seeding [98], fruit and vegetable harvesting [99,100], and weed management [101,102].
In contrast, robotic arms and gripping tools aboard a UAS are much rarer and only used for specific cases given the limitations in carrying capacity, flight time, and system balancing. Therefore, most UAS equipped with grippers and other robotic attachments will be designed for specific niche purposes. A good example comes from Konam who proposed a design that would use a robotic arm to cut ripened mangos and collect them in a net located at the base of the tree, which is work that cannot be effectively done by a ground-based system given tree heights [103]. Addressing a similar situation, Varadaramanujan and his colleagues present another UAS equipped with a robotic end-effector which can be used to pick fruits and vegetables from hard to reach areas [104]. As this category of peripheral equipment is very much still in its infancy, future applications will be addressed in the last subsection of this paper, Future Areas of Research (see Section 6.2.).

3.5. Geospatial Technology

Incorporated into almost every UAS found in agriculture is some form of geospatial technology which allows for point-specific data gathering and treatment within fields. Generally, this will include a global navigation satellite system (GNSS) receiver which is in communication with satellites. A few common GNSS systems include: GPS from the United States, Global Navigation Satellite System (GLONASS) from Russia, Galileo from the European Union, and BeiDou from China [89,105]. A GNSS is extremely useful for georeferencing images, providing better accuracy than photogrammetric methods alone [9,18,20,40,89,106]. Many projects also incorporate an inertial measurement unit (IMU) to aid in the process of georeferencing where GNSS data are temporarily unavailable [6,41,107,108]. Aside from georeferencing, another popular use for geospatial technology is for geofencing, which can be defined as an autonomous method for creating restraints on the operating area the UAS is allowed to travel [109].

4. Data Processing

4.1. Vegetation

Potentially, the most important image analysis tools available in agriculture production are vegetation indices. Vegetation indices are numerical values generated through relationships between different wavelengths of light reflected from plants [82]. Depending on the index, these values can be associated with particular properties of plant health and growth including canopy greenness, pigment composition, water stress, nutrient stress, photosynthetic potential, and diseased tissue [26,33,68]. While many common vegetation indices rely upon multispectral imaging to provide a NIR band, a number of useful indices can be calculated using strictly RGB cameras as seen in the work done by Fuentes-Peailillo and his colleagues in 2019. They compared multispectral-based vegetation indices with RGB-based indices and found that the cheaper and easier to use RGB-based indices can still provide useful information for large scale applications and simple analysis [110].
The standard and most commonly used vegetation index is the NDVI and it has been applied to numerous studies across many areas of use [25]. The equation used to calculate NDVI is as follows:
N D V I =   N I R     R N I R   +   R
where NIR represents near infrared and R represents the red color band. Generally, the NIR value comes from reflected light around the 780–800 nm range and the R value comes from reflected light around the 670–700 nm range [23,25,41]. The intensities of these variables must be measured using a multispectral or hyperspectral camera and then formatted into an NDVI map. Therefore, in a resulting NDVI map, every pixel represents a particular NDVI value. Often, these maps are visualized by associating color maps with different levels of NDVI intensity to show variations in the field. Likewise, most other vegetation indices can be visualized in a similar fashion.
Aside from NDVI, there are numerous other indices used for agricultural purposes. For example, LAI which measures leaf density per unit area and can provide information regarding the spatial variability of agricultural crops in a field [7]. Optimized soil-adjusted vegetation index (OSAVI) attempts to correct for interfering soil reflectance in NDVI and can be useful for estimating biomass and growth [7,111]. If thermal imaging tools are available, CWSI values can be generated [87,89]. With only RGB imaging available, there are still a few indices that can provide useful information such as excess red (ExR), excess green (ExG), and excess blue (ExB), which are visualized in Figure 2 [112]. The equations are as follows:
E x R = 1.4 R G
E x G = 2 G R B
E x B = 1.4 B G
where R represents red, G represents green, and B represents blue. It should be noted that ExG is certainly the most common of these indices and used to determine greenness [47,56,113].
Vegetation indices are constantly being created and updated to serve particular research needs. In addition, novel indices can also be generated by combining pre-existing indices [8,32]; therefore, it would be impossible to provide a complete list of indices currently in use.

4.2. 3D Point Clouds

Along with vegetation indices, the generation of point clouds are another very common reason to use UAS in agricultural research. The concept is fairly intuitive; a 3D point cloud is a group of points in three-dimensional space which outlines the surface of an object [114]. These points are collected by various means. Honkavaara and her colleagues used next generation automated terrain extraction software (NGATE) to generate point clouds based upon NIR images taken from a UAS [39]. Torres-Sánchez and his colleagues used a combination of a few ground control points and standard RGB images taken from their UAS along with computer vision software to generate their point clouds [78]. Potena and his colleagues developed AgriColMap, which incorporates a vegetation index map and digital surface model (DSM) which were developed using point clouds, through a combination of UAS and UGV [115]. Other methods for the collection of point clouds include using SfM algorithms (a form of photogrammetric software) [21] and RGB-D imaging [69]. From a 3D point cloud, it is possible to generate many different types of 3D models, such as DSM which are commonly used in agricultural research [14,38,39,40,78]. Likewise, multiple 3D point clouds generated over a span of time to model growth or changes can be labeled as 4D point clouds. This model is becoming increasingly prevalent in agricultural applications; thus it can be possible to monitor the growth patterns of field crops over the course of a season [116,117]. Fortunately, the modeling process is made easier through photogrammetric software such as Agisoft PhotoScan Professional (Agisoft LLC, St. Petersburg, Russia) [14,38,78] or through the use of open source libraries such as Point Cloud Library (PCL) [69,118]. Point clouds and 3D modeling have primarily been used in agriculture for the purposes of biomass estimation, 3D tree and crop modeling, and weed detection.

4.3. Data Processing

4.3.1. Image Processing

The processes involved in image processing will be different for different research goals and from the use of differing imaging paradigms. For the generation of field orthomosaics, geo-referenced images will generally be captured in a RAW format and then corrected for various distortions, interferences, and vignetting [72]. Vignetting can be especially important, as this process corrects for the nature of photographs to have a higher level of brightness in the center compared to the edges, which can lead to inconsistencies in the resulting map. All of the individual georeferenced images will then be stitched together to create the orthomosaic [7]. For multispectral and hyperspectral imaging, image processing is more complex and involves calibration and data correction steps which have to consider aspects such as atmospheric weather and noise, though there is software available to help with this process [33,80]. Radiometric calibration is required for these sensors, which often involves utilizing the empirical line method on ground targets [119]. As these sensors are sensitive to varying levels of illumination due to the changing cloud cover and time of day, light sensors have been developed to correct for these discrepancies [23].

4.3.2. Data Processing

Unmanned aircraft systems are capable of generating very large amounts of data through their sensors, but these data are generally useless without proper data processing methods. Currently, machine learning methods have risen in popularity to help make sense of the vast datasets gathered by UAS. A good example of this can be seen in the paper by Pflanz and his colleagues which contained a dataset of 25,452 sub-images for their weed mapping and classification [51]. Popular machine learning models include random forest classifiers, support vector machines, bag of visual words, and convolutional neural networks (CNNs) [29,72,120]. The latter, in particular, have become an increasingly prominent tool for data analysis in agriculture as this is considered one of the best methods for effective image processing and classification [120]. This is a form of deep learning, which relies on multiple (sometimes hundreds or thousands) of neural network layers. There have been a wide variety of CNN models developed which can be selected at will to fit the needs of the project. Some boast very high accuracy, while some trade some accuracy for speed. As CNNs will require a very large training set (many popular models are trained on millions of images) to be effective and able to learn basic shapes in images, it is recommended to borrow an existing model and re-train it on a new dataset by freezing the lower levels responsible for basic abstraction of lines and shapes, and unfreezing the higher levels of abstraction (e.g., where the lines and shapes from lower levels begin to form objects and more complex structures) to be able to develop new classification filters for any unique dataset [35,121]. Traditional statistical analysis is also commonplace in some areas of research. Common methods in this category include the use of extremum estimators, spatial statistics, linear classifiers, image statistics, and focal statistics [4,9,21,31,62].

4.3.3. Multi-Sensor Processing

Since the system aspect of UAS is what separates the terminology from simple UAVs, it follows that many systems will be composed of multiple sensors, often working simultaneously. For processing all of this data, there will often be some form of on-board computer such as a Raspberry Pi [24] or ODOIRD [122], or a ground control station (GCS); though, Luxhøj points out that disruptions in transmission between a UAS and its GCS can cause many issues [123]. For some projects, relying solely on on-board computing is not feasible either financially or computationally. For example, if a project relies heavily on data mining or machine learning, an on-board computer will generally not have the required processing power. A GCS is also useful for integration with wireless sensor network technology as discussed in the Integration and Usability subsection (see Section 5.3).

4.3.4. Autonomy

Autonomy is an important aspect of UAS technology, not only for ease of use and efficiency purposes, but also for safety reasons [109]. The innovation of geofencing tools and sense-and-avoid technology in UAS have allowed for much safer flights [124]. Unmanned aircraft system autonomy cannot be accurately classified through a binary system, but instead lies upon a spectrum with multiple levels of magnitude. Different systems to classify UAS autonomy exist, such as the 10 levels of UAV autonomy described by the Air Force Research Laboratory (OH, USA) [125]; however, UAS used in agriculture are much less complex than those developed in military labs and will generally fall into two forms of autonomy. The first form of autonomy is where the pilot will plan out the flight ahead of time and the UAS will follow a set path. The majority of agricultural research using UAS autonomy will fall into this category [17,24]. These flight paths are generated through various flight planning software such as APM Planner (version 1.1.26) used by Koh and Wich [15], Michael Oborne’s Mission Planner used by Stefanakis and his colleagues [74], or through a custom-built program as done by Chapman and colleagues [67]. The second form of UAS autonomy emerging in agricultural research is complete autonomy capable of independent decision making and flight planning based upon sensor data. In 2017, Alsalam and his colleagues developed a system called On-board Decision Making Approach which uses ultrasonic sensors and a camera communicating with a ground station to automatically detect and spray weeds in a field [122].

5. Considerations

5.1. Legal

Both UAV and UAS laws and regulating bodies can be difficult to find in many areas, as the legality has not quite caught up to the emerging technology. Depending on the location, laws can be drastically different or non-existent. In the United States, the regulatory body is the Federal Aviation Administration (FAA). For recreational purposes, the owner of the UAS must register it and apply the registration number to the drone, fly only below 400 feet in class G airspace, and keep the drone in sight at all times, among a few other safety rules that can be found on their website [126]. For commercial usage, the owner/operator must also pass a knowledge test and become an FAA-certified drone pilot [127]. For the spraying of pesticide and other hazardous chemicals, the legality in the United States is still lacking fluidity and transparency. As of 2018, it is legally possible to spray pesticide via UAS; however, the operator must go through three exemption and waiver processes to receive permission [128]. Privacy concerns are also a pressing issue surrounding drone usage; though, for agriculture, this is less of an issue as there is generally not a need to leave one’s own property [129]. More legal information can be found online as there are many organizations and websites which aim to help provide guidance in this area based upon location of use [130,131,132].

5.2. Economic

Economic aspects are one of the primary motivations for agricultural research involving UAS. Many research projects in agriculture intend to eventually provide support for farmers and the food-generating community as a whole, so the economic impact of any advancement in technology or methodology should be evaluated. Unmanned aircraft systems aim to provide easier means for monitoring and treating fields, aligned with the precision agriculture paradigm [109], but there is a financial trade-off of which to be aware. Many of these systems for continuous usage can be expensive, costing tens of thousands of dollars and are not realistic for many farms [17,77]. While the UAS platform itself can be very affordable, the specialized equipment and cameras are not. For example, the DJI Mavic 2 Pro, a popular consumer UAV fitted with a high-resolution RGB camera, costs around (US) $1500, whereas, to add a hyperspectral camera built for aerial usage could cost over (US) $10,000 [80,133]. In terms of aerial sensing platforms, a study by Matese and his colleagues found that farms smaller than 5 hectares and larger than 50 hectares would not benefit from using UAS and that satellite or aircraft sensing would be more economical [9]. However, for routine scouting and monitoring of fields that cannot be done via satellite, UAS have been shown to be a potentially economically viable replacement for hired laborers [33,34]. Unfortunately, as most of this research is still in the proof-of-concept stage, there is little hard evidence of the cost trade-offs of using UAS within agriculture [24]. As the technology develops and extension and outreach centers are able to provide farmers with the knowledge to operate these tools, the cost-prohibiting nature of these systems is expected to decrease, and many researchers are optimistic about the future [24,77].

5.3. Integration and Usability

Unmanned aircraft system integration with other sensing systems is a key element in effective precision farming. Many modern farms and nurseries have begun to incorporate ground-based sensors in the field to monitor various aspects such as soil health, pests, irrigation, and environment [134,135]. These sensing systems are usually wireless and connected over a wireless sensor network (WSN) and communicate with a base station [3,136]. Therefore, it is possible for a UAS to be treated as another part of the existing WSN. In 2015, Kale and her colleagues applied this idea to pesticide spraying by developing a framework for UAS to communicate with ground-based sensors over the WSN to reduce chemical usage and drift [64]. In 2018, Moribe and his colleagues applied this idea by integrating infrared thermometer sensors on the ground and onboard a UAS, and developed a novel communication protocol between them for the WSN [137]. Furthermore, Uddin and his colleagues built a robust crop health monitoring system through integration of ground-based wireless sensors and UAVs in 2018 [138].
Another crucial aspect of this emerging technology is the ability to translate the theory into practice. Much of the topics covered previously in this review are of a technical nature, and it is not expected that many farmers will learn the nuances of, for example, equipping and calibrating hyperspectral imaging capabilities on a consumer-grade UAV for the purpose of disease detection in their fields. Therefore, streamlining these concepts into complete, easy-to-use packages with support for WSNs is of high importance to see any real application of this technology.

6. Discussion

The goal of this review article was to provide a concise, but thorough, survey of the current usage of UAVs within the domain of agriculture. Within the above text, there is information regarding specific applications, the individual physical components of the systems and how they work together, how data are gathered and processed through them, as well as a few minor considerations of which to be aware. With this knowledge, it will be beneficial to understand the advantages and limitations of this technology.

6.1. Advantages and Limitations

Unmanned aircraft system technology is able to provide us with very high spatial resolution at the cost of temporal resolution. It is capable of traversing difficult terrain; however, many platforms suffer from short flight times, especially with large payloads. Likewise, UAS efficacy can be very dependent upon weather conditions. A large advantage of UAS is the ability to fly in different trajectories and incorporate depth sensors in order to create accurate 3D models and maps; though, it should be noted that these photogrammetric processing times can be very extensive and require a great deal of computing time when using high-resolution images. As well, UAS are capable of physically interacting with the environment, performing tasks such as sample collection and chemical spraying. Similarly, the UAS can be connected into an existing WSN for an integrated approach to precision farming. However, much of this sensing technology and software comes with a large upfront cost. Therefore, due care must be taken in calculating the economic trade-off of using UAS over other options.

6.2. Future Areas of Research

Hopefully, it is clear that the potential of UAS in agriculture is very high. This review only brushes the surface of all of the possibilities of these systems. As stated previously, there are many applications where research is lacking or non-existent in this domain. For example, the development of agricultural image databases would provide a massive aid in the creation of machine vision algorithms to run from UAS, as many researchers find themselves having to create their own datasets which can be very time consuming and costly [120]. Likewise, a centralized source for the collection of which hyperspectral bands to monitor for particular diseases or stresses could save researchers time through derivative experiments. It would be worth investigating the potential of applying data mining methods to the collection of these hyperspectral readings of light reflected from plants and soil, as there would be far too much data for traditional inference. The field of deep learning is also currently a very hot topic in academia, and many models applied to agriculture are only in their infancy. Therefore, it should not be difficult to find novel areas of research within deep learning or machine learning as a whole. It should be obvious from living in the modern era that agriculture will see a more intimate relationship with computer science over the next few decades, and researchers should be encouraged to collaborate more between these disciplines. In fact, the authors of this article are currently involved in a cross-disciplinary project between computer science and agriculture which seeks to build a UAS with the capability of automatic biotic and abiotic stress detection of ornamental crops in a nursery setting, as well as the ability to create 3D models of trees to monitor growth patterns and a robotic arm for taking tissue samples. A final notable area of future research is within the discipline of animal science. From reviewing almost 200 papers and conference proceedings, a very small minority had any focus in applying this technology to livestock. It would seem that UAS with sensors such as thermal cameras could potentially help monitor the health of herds of cattle or even aid in discovering lost or trapped individuals. Regardless of particular applications, the end goal of the majority of this research will be the creation of a complete and integrated method for aerial monitoring and treatment of various anomalies in farm settings with limited human intervention. This is the future of precision farming and likely the most effective way to conduct agriculture in the 21st century.

Author Contributions

Writing—original draft preparation, S.C.H.; writing—review and editing, F.B.-G; funding—received, F.B.-G.

Funding

This work was supported by the National Institute of Food and Agriculture, United States Department of Agriculture Capacity Building grant, under award number 2019-38821-29062.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhang, N.; Wang, M.; Wang, N. Precision agriculture—A worldwide overview. Comput. Electron. Agric. 2002, 36, 113–132. [Google Scholar] [CrossRef]
  2. Tokekar, P.; Vander Hook, J.; Mulla, D.; Isler, V. Sensor planning for a symbiotic UAV and UGV system for precision agriculture. IEEE Trans. Robot. 2016, 32, 1498–1511. [Google Scholar] [CrossRef]
  3. Kumar, S.A.; Ilango, P. The impact of wireless sensor network in the field of precision agriculture: A review. Wirel. Pers. Commun. 2018, 98, 685–698. [Google Scholar] [CrossRef]
  4. Murugan, D.; Garg, A.; Singh, D. Development of an adaptive approach for precision agriculture monitoring with drone and satellite data. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 5322–5328. [Google Scholar] [CrossRef]
  5. Kussul, N.; Lavreniuk, M.; Skakun, S.; Shelestov, A. Deep learning classification of land cover and crop types using remote sensing data. IEEE Geosci. Remote Sens. Lett. 2017, 14, 778–782. [Google Scholar] [CrossRef]
  6. Viljanen, N.; Honkavaara, E.; Näsi, R.; Hakala, T.; Niemeläinen, O.; Kaivosoja, J. A novel machine learning method for estimating biomass of grass swards using a photogrammetric canopy height model, images and vegetation indices captured by a drone. Agriculture 2018, 8, 70. [Google Scholar] [CrossRef]
  7. Marino, S.; Alvino, A. Detection of spatial and temporal variability of wheat cultivars by high-resolution vegetation indices. Agronomy 2019, 9, 226. [Google Scholar] [CrossRef]
  8. Gevaert, C.M.; Tang, J.; Suomalainen, J.; Kooistra, L. Combining hyperspectral UAV and multispectral FORMOSAT-2 imagery for precision agriculture applications. In Proceedings of the Workshop on Hyperspectral Image and Signal Processing, Lausanne, Switzerland, 24–27 June 2014. [Google Scholar]
  9. Matese, A.; Toscano, P.; Di Gennaro, S.F.; Genesio, L.; Vaccari, F.P.; Primicerio, J.; Belli, C.; Zaldei, A.; Bianconi, R.; Gioli, B. Intercomparison of UAV, aircraft and satellite remote sensing platforms for precision viticulture. Remote Sens. 2015, 7, 2971–2990. [Google Scholar] [CrossRef]
  10. Canis, B. Unmanned aircraft systems (UAS): Commercial outlook for a new industry. Congr. Res. Serv. Rep. 2015, 7-5700, R44192. [Google Scholar]
  11. Gupta, S.G.; Ghonge, M.M.; Jawandhiya, P.M. Review of unmanned aircraft system (UAS). Int. J. Adv. Res. Comput. Eng. Technol. 2013, 2, 1646–1658. [Google Scholar] [CrossRef]
  12. Zhang, C.; Kovacs, J.M. The application of small unmanned aerial systems for precision agriculture: A review. Precis. Agric. 2012, 13, 693–712. [Google Scholar] [CrossRef]
  13. Hunt, E.R.; Daughtry, C.S.T.; Mirsky, S.B.; Hively, W.D. Remote sensing with simulated unmanned aircraft imagery for precision agriculture applications. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 4566–4571. [Google Scholar] [CrossRef]
  14. Chang, A.; Jung, J.; Maeda, M.M.; Landivar, J. Crop height monitoring with digital imagery from Unmanned Aerial System (UAS). Comput. Electron. Agric. 2017, 141, 232–237. [Google Scholar] [CrossRef]
  15. Koh, L.P.; Wich, S.A. Dawn of drone ecology: Low-cost autonomous aerial vehicles for conservation. Trop. Conserv. Sci. 2012, 5, 121–132. [Google Scholar] [CrossRef]
  16. Gevaert, C.M.; Suomalainen, J.; Tang, J.; Kooistra, L. Generation of spectral-temporal response surfaces by combining multispectral satellite and hyperspectral UAV imagery for precision agriculture applications. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 3140–3146. [Google Scholar] [CrossRef]
  17. Primicerio, J.; Di Gennaro, S.F.; Fiorillo, E.; Genesio, L.; Lugato, E.; Matese, A.; Vaccari, F.P. A flexible unmanned aerial vehicle for precision agriculture. Precis. Agric. 2012, 13, 517–523. [Google Scholar] [CrossRef]
  18. Bachmann, F.; Herbst, R.; Gebbers, R.; Hafner, V.V. Micro UAV based georeferenced orthophoto generation in VIS + NIR for precision agriculture. ISPRS Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2013, XL-1/W2, 11–16. [Google Scholar] [CrossRef]
  19. Navia, J.; Mondragon, I.; Patino, D.; Colorado, J. Multispectral mapping in agriculture: Terrain mosaic using an autonomous quadcopter UAV. In Proceedings of the IEEE 2016 International Conference on Unmanned Aircraft Systems, ICUAS 2016, Arlington, VA, USA, 7–10 June 2016; pp. 1351–1358. [Google Scholar]
  20. Nolan, A.P.; Park, S.; Fuentes, S.; Ryu, D.; Chung, H. Automated detection and segmentation of vine rows using high resolution UAS imagery in a commercial vineyard. In Proceedings of the 21st International Congress on Modelling and Simulation, Gold Coast, Australia, 29 November–4 December 2015; pp. 1406–1412. [Google Scholar]
  21. Surový, P.; Almeida Ribeiro, N.; Panagiotidis, D. Estimation of positions and heights from UAV-sensed imagery in tree plantations in agrosilvopastoral systems. Int. J. Remote Sens. 2018, 39, 4786–4800. [Google Scholar] [CrossRef]
  22. Rokhmana, C.A. The potential of UAV-based remote sensing for supporting precision agriculture in Indonesia. Procedia Environ. Sci. 2015, 24, 245–253. [Google Scholar] [CrossRef]
  23. Roth, L.; Streit, B. Predicting cover crop biomass by lightweight UAS-based RGB and NIR photography: An applied photogrammetric approach. Precis. Agric. 2017, 19, 1–22. [Google Scholar] [CrossRef]
  24. Katsigiannis, P.; Misopolinos, L.; Liakopoulos, V.; Alexandridis, T.K.; Zalidis, G. An autonomous multi-sensor UAV system for reduced-input precision agriculture applications. In Proceedings of the IEEE 24th Mediterranean Conference on Control and Automation, MED 2016, Athens, Greece, 21–24 June 2016; pp. 60–64. [Google Scholar]
  25. Gago, J.; Douthe, C.; Coopman, R.E.; Gallego, P.P.; Ribas-Carbo, M.; Flexas, J.; Escalona, J.; Medrano, H. UAVs challenge to assess water stress for sustainable agriculture. Agric. Water Manag. 2015, 153, 9–19. [Google Scholar] [CrossRef]
  26. Zaman-Allah, M.; Vergara, O.; Araus, J.L.; Tarekegne, A.; Magorokosho, C.; Zarco-Tejada, P.J.; Hornero, A.; Albà, A.H.; Das, B.; Craufurd, P.; et al. Unmanned aerial platform-based multi-spectral imaging for field phenotyping of maize. Plant Methods 2015, 11, 35. [Google Scholar] [CrossRef] [PubMed]
  27. Cilia, C.; Panigada, C.; Rossini, M.; Meroni, M.; Busetto, L.; Amaducci, S.; Boschetti, M.; Picchi, V.; Colombo, R. Nitrogen status assessment for variable rate fertilization in maize through hyperspectral imagery. Remote Sens. 2014, 6, 6549–6565. [Google Scholar] [CrossRef]
  28. Li, J.; Zhang, F.; Qian, X.; Zhu, Y.; Shen, G. Quantification of rice canopy nitrogen balance index with digital imagery from unmanned aerial vehicle. Remote Sens. Lett. 2015, 6, 183–189. [Google Scholar] [CrossRef]
  29. Garcia-Ruiz, F.; Sankaran, S.; Maja, J.M.; Lee, W.S.; Rasmussen, J.; Ehsani, R. Comparison of two aerial imaging platforms for identification of Huanglongbing-infected citrus trees. Comput. Electron. Agric. 2013, 91, 106–115. [Google Scholar] [CrossRef]
  30. Al-Saddik, H.; Simon, J.C.; Brousse, O.; Cointault, F. Multispectral band selection for imaging sensor design for vineyard disease detection: Case of Flavescence Dorée. Adv. Anim. Biosci. 2017, 8, 150–155. [Google Scholar] [CrossRef]
  31. De Castro, A.I.; Ehsani, R.; Ploetz, R.C.; Crane, J.H.; Buchanon, S. Detection of laurel wilt disease in avocado using low altitude aerial imaging. PLoS ONE 2015, 10, e0124642. [Google Scholar] [CrossRef] [PubMed]
  32. Mattupalli, C.; Moffet, C.A.; Shah, K.N.; Young, C.A. Supervised classification of RGB Aerial imagery to evaluate the impact of a root rot disease. Remote Sens. 2018, 10, 917. [Google Scholar] [CrossRef]
  33. Zhang, D.; Zhou, X.; Zhang, J.; Lan, Y.; Xu, C.; Liang, D. Detection of rice sheath blight using an unmanned aerial system with high-resolution color and multispectral imaging. PLoS ONE 2018, 13, e0187470. [Google Scholar] [CrossRef]
  34. Kalischuk, M.; Paret, M.L.; Freeman, J.H.; Raj, D.; Da Silva, S.; Eubanks, S.; Wiggins, D.J.; Lollar, M.; Marois, J.J.; Mellinger, H.C.; et al. An improved crop scouting technique incorporating unmanned aerial vehicle–assisted multispectral crop imaging into conventional scouting practice for gummy stem blight in watermelon. Plant Dis. 2019, 103, 1642–1650. [Google Scholar] [CrossRef]
  35. Sladojevic, S.; Arsenovic, M.; Anderla, A.; Culibrk, D.; Stefanovic, D. Deep neural networks based recognition of plant diseases by leaf image classification. Comput. Intell. Neurosci. 2016, 2016, 3289801. [Google Scholar] [CrossRef] [PubMed]
  36. Sabrol, H.; Satish, K. Tomato plant disease classification in digital images using classification tree. In Proceedings of the IEEE International Conference on Communication and Signal Processing, ICCSP 2016, Melmaruvathur, India, 6–8 April 2016; pp. 1242–1246. [Google Scholar]
  37. Dhaware, C.G.; Wanjale, K.H. A modern approach for plant leaf disease classification which depends on leaf image processing. In Proceedings of the 2017 International Conference on Computer Communication and Informatics, ICCCI 2017, Coimbatore, India, 5–7 January 2017. [Google Scholar]
  38. Grüner, E.; Astor, T.; Wachendorf, M. Biomass prediction of heterogeneous temperate grasslands using an SfM approach based on UAV imaging. Agronomy 2019, 9, 54. [Google Scholar] [CrossRef]
  39. Honkavaara, E.; Kaivosoja, J.; Mäkynen, J.; Pellikka, I.; Pesonen, L.; Saari, H.; Salo, H.; Hakala, T.; Marklelin, L.; Rosnell, T. Hyperspectral reflectance signatures and point clouds for precision agriculture by light weight UAV imaging system. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2012, I–7, 353–358. [Google Scholar] [CrossRef]
  40. Honkavaara, E.; Saari, H.; Kaivosoja, J.; Pölönen, I.; Hakala, T.; Litkey, P.; Mäkynen, J.; Pesonen, L. Processing and assessment of spectrometric, stereoscopic imagery collected using a lightweight UAV spectral camera for precision agriculture. Remote Sens. 2013, 5, 5006–5039. [Google Scholar] [CrossRef]
  41. Geipel, J.; Link, J.; Wirwahn, J.; Claupein, W. A programmable aerial multispectral camera system for in-season crop biomass and nitrogen content estimation. Agriculture 2016, 6, 4. [Google Scholar] [CrossRef]
  42. Aldana-Jague, E.; Heckrath, G.; Macdonald, A.; van Wesemael, B.; Van Oost, K. UAS-based soil carbon mapping using VIS-NIR (480–1000 nm) multi-spectral imaging: Potential and limitations. Geoderma 2016, 275, 55–66. [Google Scholar] [CrossRef]
  43. Thilakarathna, M.; Raizada, M. Challenges in using precision agriculture to optimize symbiotic nitrogen fixation in legumes: Progress, limitations, and future improvements needed in diagnostic testing. Agronomy 2018, 8, 78. [Google Scholar] [CrossRef]
  44. Berger, K.; Atzberger, C.; Danner, M.; D’Urso, G.; Mauser, W.; Vuolo, F.; Hank, T. Evaluation of the PROSAIL model capabilities for future hyperspectral model environments: A review study. Remote Sens. 2018, 10, 85. [Google Scholar] [CrossRef]
  45. Duan, S.B.; Li, Z.L.; Wu, H.; Tang, B.H.; Ma, L.; Zhao, E.; Li, C. Inversion of the PROSAIL model to estimate leaf area index of maize, potato, and sunflower fields from unmanned aerial vehicle hyperspectral data. Int. J. Appl. Earth Obs. Geoinf. 2014, 26, 12–20. [Google Scholar] [CrossRef]
  46. Verger, A.; Vigneau, N.; Chéron, C.; Gilliot, J.M.; Comar, A.; Baret, F. Green area index from an unmanned aerial system over wheat and rapeseed crops. Remote Sens. Environ. 2014, 152, 654–664. [Google Scholar] [CrossRef]
  47. Rasmussen, J.; Nielsen, J.; Garcia-Ruiz, F.; Christensen, S.; Streibig, J.C. Potential uses of small unmanned aircraft systems (UAS) in weed research. Weed Res. 2013, 53, 242–248. [Google Scholar] [CrossRef]
  48. Sandler, H. Weed management in cranberries: A historical perspective and a look to the future. Agriculture 2018, 8, 138. [Google Scholar] [CrossRef]
  49. Gómez-Candón, D.; De Castro, A.I.; López-Granados, F. Assessing the accuracy of mosaics from unmanned aerial vehicle (UAV) imagery for precision agriculture purposes in wheat. Precis. Agric. 2014, 15, 44–56. [Google Scholar] [CrossRef]
  50. Mink, R.; Dutta, A.; Peteinatos, G.; Sökefeld, M.; Engels, J.; Hahn, M.; Gerhards, R. Multi-temporal site-specific weed control of Cirsium arvense (L.) Scop. and Rumex crispus L. in maize and sugar beet using unmanned aerial vehicle based mapping. Agriculture 2018, 8, 65. [Google Scholar] [CrossRef]
  51. Pflanz, M.; Nordmeyer, H.; Schirrmann, M. Weed mapping with UAS imagery and a bag of visual words based image classifier. Remote Sens. 2018, 10, 1530. [Google Scholar] [CrossRef]
  52. Bah, M.D.; Hafiane, A.; Canals, R. Weeds detection in UAV imagery using SLIC and the hough transform. In Proceedings of the 7th International Conference on Image Processing Theory, Tools and Applications, IPTA 2017, Montreal, QC, Canada, 28 November–1 December 2017; Volume 2018, pp. 1–6. [Google Scholar]
  53. Rasmussen, J.; Nielsen, J.; Streibig, J.C.; Jensen, J.E.; Pedersen, K.S.; Olsen, S.I. Pre-harvest weed mapping of Cirsium arvense in wheat and barley with off-the-shelf UAVs. Precis. Agric. 2018, 20, 983–999. [Google Scholar] [CrossRef]
  54. Zortea, M.; Macedo, M.M.G.; Mattos, A.B.; Ruga, B.C.; Gemignani, B.H. Automatic citrus tree detection from UAV images based on convolutional neural networks. In Proceedings of the 31th Sibgrap/WIA—Conference on Graphics, Patterns and Images, SIBGRAPI’18, Foz do Iguacu, Brazil, 29 October–1 November 2018. [Google Scholar]
  55. She, Y.; Ehsani, R.; Robbins, J.; Leiva, J.N.; Owen, J. Applications of small UAV systems for tree and nursery inventory management. In Proceedings of the 12th International Conference on Precision Agriculture (ICPA), Sacramento, CA, USA, 20–23 July 2014. [Google Scholar]
  56. She, Y.; Ehsani, R.; Robbins, J.; Leiva, J.N.; Owen, J. Applications of high-resolution imaging for open field container nursery counting. Remote Sens. 2018, 10, 12. [Google Scholar] [CrossRef]
  57. Rahnemoonfar, M.; Sheppard, C. Deep count: Fruit counting based on deep simulated learning. Sensors 2017, 17, 905. [Google Scholar] [CrossRef] [PubMed]
  58. Rahnemoonfar, M.; Dobbs, D.; Yari, M.; Starek, M.J. DisCountNet: Discriminating and counting network for real-time counting and localization of sparse objects in high-resolution UAV imagery. Remote Sens. 2019, 11, 1128. [Google Scholar] [CrossRef]
  59. Pederi, Y.A.; Cheporniuk, H.S. Unmanned aerial vehicles and new technological methods of monitoring and crop protection in precision agriculture. In Proceedings of the 2015 IEEE 3rd International Conference Actual Problems of Unmanned Aerial Vehicles Developments, APUAVD 2015—Proceedings, Kiev, Ukraine, 13–15 October 2015; pp. 298–301. [Google Scholar]
  60. Xiao, Q.; Xin, F.; Lou, Z.; Zhou, T.; Wang, G.; Han, X.; Lan, Y.; Fu, W. Effect of aviation spray adjuvants on defoliant droplet deposition and cotton defoliation efficacy sprayed by unmanned aerial vehicles. Agronomy 2019, 9, 217. [Google Scholar] [CrossRef]
  61. Zhu, H.; Li, H.; Zhang, C.; Li, J.; Zhang, H. Performance characterization of the UAV chemical application based on CFD simulation. Agronomy 2019, 9, 308. [Google Scholar] [CrossRef]
  62. Lou, Z.; Xin, F.; Han, X.; Lan, Y.; Duan, T.; Fu, W. Effect of Unmanned Aerial Vehicle Flight Height on Droplet Distribution, Drift and Control of Cotton Aphids and Spider Mites. Agronomy 2018, 8, 187. [Google Scholar] [CrossRef]
  63. Myers, D.; Ross, C.; Liu, B.; Ave, G.; Poly, C.; Obispo, S.L. A review of unmanned aircraft system (UAS) applications for agriculture. In Proceedings of the 2015 ASABE Annual International Meeting, New Orleans, LA, USA, 26-29 July 2015. [Google Scholar]
  64. Kale, S.D.; Khandagale, S.V.; Gaikwad, S.S.; Narve, S.S.; Gangal, P.V. International journal of advanced research in computer science and software engineering agriculture drone for spraying fertilizer and pesticides. Int. J. Adv. Res. Comput. Sci. Softw. Eng. 2015, 5, 804–807. [Google Scholar]
  65. Mogili, U.R.; Deepak, B.B.V.L. Review on application of drone systems in precision agriculture. Procedia Comput. Sci. 2018, 133, 502–509. [Google Scholar] [CrossRef]
  66. Deery, D.; Jimenez-Berni, J.; Jones, H.; Sirault, X.; Furbank, R. Proximal remote sensing buggies and potential applications for field-based phenotyping. Agronomy 2014, 5, 349–379. [Google Scholar] [CrossRef]
  67. Chapman, S.; Merz, T.; Chan, A.; Jackway, P.; Hrabar, S.; Dreccer, M.; Holland, E.; Zheng, B.; Ling, T.; Jimenez-Berni, J. Pheno-copter: A low-altitude, autonomous remote-sensing robotic helicopter for high-throughput field-based phenotyping. Agronomy 2014, 4, 279–301. [Google Scholar] [CrossRef]
  68. Patrick, A.; Pelham, S.; Culbreath, A.; Corely Holbrook, C.; De Godoy, I.J.; Li, C. High throughput phenotyping of tomato spot wilt disease in peanuts using unmanned aerial systems and multispectral imaging. IEEE Instrum. Meas. Mag. 2017, 20, 4–12. [Google Scholar] [CrossRef]
  69. Tripicchio, P.; Satler, M.; Dabisias, G.; Ruffaldi, E.; Avizzano, C.A. Towards smart farming and sustainable agriculture with drones. In Proceedings of the 11th International Conference on Intelligent Environments, Prague, Czech Republic, 15–17 July 2015; pp. 140–143. [Google Scholar]
  70. Pobkrut, T.; Eamsa-Ard, T.; Kerdcharoen, T. Sensor drone for aerial odor mapping for agriculture and security services. In Proceedings of the 13th IEEE International Conference on Electrical Engineering/Electronics, Computer, Telecommunications and Information Technology, Chiang Mai, Thailand, 28 June–1 July 2016. [Google Scholar]
  71. Rangel, R.K. Development of an UAVS distribution tools for pest’s biological control “Bug Bombs!”. In Proceedings of the IEEE Aerospace Conference Proceedings, Big Sky, MT, USA, 5–12 March 2016. [Google Scholar]
  72. Tewes, A.; Schellberg, J. Towards remote estimation of radiation use efficiency in maize using UAV-based low-cost camera imagery. Agronomy 2018, 8, 16. [Google Scholar] [CrossRef]
  73. Sankaran, S.; Khot, L.R.; Espinoza, C.Z.; Jarolmasjed, S.; Sathuvalli, V.R.; Vandemark, G.J.; Miklas, P.N.; Carter, A.H.; Pumphrey, M.O.; Knowles, R.R.N.; et al. Low-altitude, high-resolution aerial imaging systems for row and field crop phenotyping: A review. Eur. J. Agron. 2015, 70, 112–123. [Google Scholar] [CrossRef]
  74. Stefanakis, D.; Hatzopoulos, J.N.; Margaris, N. Creation of a remote sensing unmanned aerial system (UAS) for precision agriculture and related mapping applications. In Proceedings of the ASPRS 2013 Annual Conference, Baltimore, MA, USA, 24–28 March 2013. [Google Scholar]
  75. Hogan, S.D.; Kelly, M.; Stark, B.; Chen, Y.Q. Unmanned aerial systems for agriculture and natural resources. Calif. Agric. 2017, 71, 5–14. [Google Scholar] [CrossRef]
  76. Yallappa, D.; Veerangouda, M.; Maski, D.; Palled, V.; Bheemanna, M. Development and evaluation of drone mounted sprayer for pesticide applications to crops. In Proceedings of the 2017 IEEE Global Humanitarian Technology Conference (GHTC), San Jose, CA, USA, 19–22 October 2017. [Google Scholar]
  77. Reinecke, M.; Prinsloo, T. The influence of drone monitoring on crop health and harvest size. In Proceedings of the 2017 1st International Conference on Next Generation Computing Applications (NextComp), Port Louis, Mauritius, 19–21 July 2017. [Google Scholar]
  78. Torres-Sánchez, J.; de Castro, A.I.; Peña, J.M.; Jiménez-Brenes, F.M.; Arquero, O.; Lovera, M.; López-Granados, F. Mapping the 3D structure of almond trees using UAV acquired photogrammetric point clouds and object-based image analysis. Biosyst. Eng. 2018, 176, 172–184. [Google Scholar] [CrossRef]
  79. Iqbal, F.; Lucieer, A.; Barry, K. Simplified radiometric calibration for UAS-mounted multispectral sensor. Eur. J. Remote Sens. 2018, 51, 301–313. [Google Scholar] [CrossRef]
  80. Adão, T.; Hruška, J.; Pádua, L.; Bessa, J.; Peres, E.; Morais, R.; Sousa, J.J. Hyperspectral imaging: A review on UAV-based sensors, data processing and applications for agriculture and forestry. Remote Sens. 2017, 9, 1110. [Google Scholar] [CrossRef]
  81. Mahajan, U.; Bundel, B.R. Drones for normalized difference vegetation index ( NDVI ), to estimate crop health for precision agriculture: A cheaper alternative for spatial satellite sensors. In Proceedings of the International Conference on Innovative Research in Agriculture, Food Science, Forestry, Horticulture, Aquaculture, Animal Sciences, Biodiversity, Ecological Sciences and Climate Change, New Delhi, India, 22 October 2016; pp. 38–41. [Google Scholar]
  82. Lowe, A.; Harrison, N.; French, A.P. Hyperspectral image analysis techniques for the detection and classification of the early onset of plant disease and stress. Plant Methods 2017, 13, 80. [Google Scholar] [CrossRef] [PubMed]
  83. Proctor, C.; He, Y. Workflow for building a hyperspectral UAV: Challenges and opportunities. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. ISPRS Arch. 2015, 40, 415–419. [Google Scholar] [CrossRef]
  84. Nackaerts, K.; Everaerts, J.; Michiels, B.; Holmlund, C.; Saari, H. Evaluation of a lightweigth UAS-prototype for hyperspectral imaging. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2010, XXXVIII, 478–483. [Google Scholar]
  85. Saari, H.; Akujärvi, A.; Holmlund, C.; Ojanen, H.; Kaivosoja, J.; Nissinen, A.; Niemeläinen, O. Visible, very near IR and short wave IR hyperspectral drone imaging system for agriculture and natural water applicationS. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. ISPRS Arch. 2017, 42, 165–170. [Google Scholar] [CrossRef]
  86. Honkavaara, E.; Hakala, T.; Markelin, L.; Jaakkola, A.; Saari, H.; Ojanen, H.; Pölönen, I.; Tuominen, S.; Näsi, R.; Rosnell, T.; et al. Autonomous hyperspectral UAS photogrammetry for environmental monitoring applications. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. ISPRS Arch. 2014, 40, 155–159. [Google Scholar] [CrossRef]
  87. Costa, J.M.; Grant, O.M.; Chaves, M.M. Thermography to explore plant-environment interactions. J. Exp. Bot. 2013, 64, 3937–3949. [Google Scholar] [CrossRef]
  88. Granum, E.; Pérez-Bueno, M.L.; Calderón, C.E.; Ramos, C.; de Vicente, A.; Cazorla, F.M.; Barón, M. Metabolic responses of avocado plants to stress induced by Rosellinia necatrix analysed by fluorescence and thermal imaging. Eur. J. Plant Pathol. 2015, 142, 625–632. [Google Scholar] [CrossRef]
  89. Ribeiro-Gomes, K.; Hernández-López, D.; Ortega, J.F.; Ballesteros, R.; Poblete, T.; Moreno, M.A. Uncooled thermal camera calibration and optimization of the photogrammetry process for UAV applications in agriculture. Sensors 2017, 17, 2173. [Google Scholar] [CrossRef] [PubMed]
  90. Vit, A.; Shani, G. Comparing RGB-D sensors for close range outdoor agricultural phenotyping. Sensors 2018, 18, 4413. [Google Scholar] [CrossRef] [PubMed]
  91. Wang, W.; Li, C. Size estimation of sweet onions using consumer-grade RGB-depth sensor. J. Food Eng. 2014, 142, 153–162. [Google Scholar] [CrossRef]
  92. Saha, A.K.; Saha, J.; Ray, R.; Sircar, S.; Dutta, S.; Chattopadhyay, S.P.; Saha, H.N. IOT-based drone for improvement of crop quality in agricultural field. In Proceedings of the 2018 IEEE 8th Annual Computing and Communication Workshop and Conference, Las Vegas, NV, USA, 8–10 January 2018; pp. 612–615. [Google Scholar]
  93. Zollhöfer, M.; Stotko, P.; Görlitz, A.; Theobalt, C.; Nießner, M.; Klein, R.; Kolb, A. State of the Art on 3D Reconstruction with RGB-D Cameras; Wiley: Hoboken, NJ, USA, 2018. [Google Scholar]
  94. Andújar, D.; Dorado, J.; Fernández-Quintanilla, C.; Ribeiro, A. An approach to the use of depth cameras for weed volume estimation. Sensors 2016, 16, 1–11. [Google Scholar] [CrossRef] [PubMed]
  95. Stark, B.; Rider, S.; Chen, Y.Q. Optimal pest management by networked unmanned cropdusters in precision agriculture: A cyber-physical system approach. In Proceedings of the 2nd IFAC Workshop on Research, Education and Development of Unmanned Aerial Systems, IFAC, Compiegne, France, 20–22 November 2013; pp. 296–302. [Google Scholar]
  96. Salama, S.; Hajjaj, H. Review of agriculture robotics: Practicality and feasibility. In Proceedings of the 2016 IEEE International Symposium on Robotics and Intelligent Sensors (IRIS), Tokyo, Japan, 17–20 December 2016; pp. 194–198. [Google Scholar]
  97. Gealy, D.V.; McKinley, S.; Guo, M.; Miller, L.; Vougioukas, S.; Viers, J.; Carpin, S.; Goldberg, K. DATE: A handheld co-robotic device for automated tuning of emitters to enable precision irrigation. In Proceedings of the IEEE International Conference on Automation Science and Engineering, Fort Worth, TX, USA, 21–24 August 2016; pp. 922–927. [Google Scholar]
  98. Thatshayini, D. FPGA realization of fuzzy based robotic manipulator for agriculture applications. In Proceedings of the 2019 1st International Conference on Innovations in Information and Communication Technology (ICIICT), Chennai, India, 25–26 April 2019. [Google Scholar]
  99. Font, D.; Pallejà, T.; Tresanchez, M.; Runcan, D.; Moreno, J.; Martínez, D.; Teixidó, M.; Palacín, J. A proposal for automatic fruit harvesting by combining a low cost stereovision camera and a robotic arm. Sensors 2014, 14, 11557–11579. [Google Scholar] [CrossRef]
  100. Roshanianfard, A.; Noguchi, N.; Kamata, T. Design and performance of a robotic arm for farm use. Int. J. Agric. Biol. Eng. 2019, 12, 146–158. [Google Scholar] [CrossRef]
  101. Sabanci, K.; Aydin, C. Smart robotic weed control system for sugar beet. J. Agric. Sci. Technol. 2017, 19, 73–83. [Google Scholar]
  102. Shamshiri, R.R.; Weltzien, C.; Hameed, I.A.; Yule, I.J.; Grift, T.E.; Balasundram, S.K.; Pitonakova, L.; Ahmad, D.; Chowdhary, G. Research and development in agricultural robotics: A perspective of digital farming. Int. J. Agric. Biol. Eng. 2018, 11, 1–14. [Google Scholar] [CrossRef]
  103. Konam, S. Agricultural aid for mango cutting (AAM). In Proceedings of the IEEE International Conference on Advances in Computing, Communications and Informatics, New Delhi, India, 24–27 September 2014; pp. 1520–1524. [Google Scholar]
  104. Varadaramanujan, S.; Sreenivasa, S.; Pasupathy, P.; Calastawad, S.; Morris, M.; Tosunoglu, S. Design of a drone with a robotic end-effector. In Proceedings of the 30th Florida Conference on Recent Advances in Robotics, Boca Raton, FL, USA, 11–12 May 2017. [Google Scholar]
  105. Guo, J.; Li, X.; Li, Z.; Hu, L.; Yang, G.; Zhao, C.; Fairbairn, D.; Watson, D.; Ge, M. Multi-GNSS precise point positioning for precision agriculture. Precis. Agric. 2018, 19, 895–911. [Google Scholar] [CrossRef] [Green Version]
  106. Pauly, K. Applying conventional vegetation vigor indices to UAS-derived orthomosaics: Issues and considerations. In Proceedings of the 12th International Conference for Precision Agriculture, Sacramento, CA, USA, 20–23 July 2014. [Google Scholar]
  107. Kaivosoja, J.; Pesonen, L.; Kleemola, J.; Pölönen, I.; Salo, H.; Honkavaara, E.; Saari, H.; Mäkynen, J.; Rajala, A. A case study of a precision fertilizer application task generation for wheat based on classified hyperspectral data from UAV combined with farm history data. In Proceedings of the SPIE—The International Society for Optics and Photonics SPIE, Dresden, Germany, 23–26 September 2013. [Google Scholar]
  108. Fuertes, F.C.; Wilhelm, L.; Porté-Agel, F. Multirotor UAV-based platform for the measurement of atmospheric turbulence: Validation and signature detection of tip vortices of wind turbine blades. J. Atmos. Ocean. Technol. 2019, 36, 941–955. [Google Scholar] [CrossRef]
  109. Atkins, E.M. Autonomy as an enabler of economically-viable, beyond-line-of-sight, low-altitude UAS applications with acceptable risk. In Proceedings of the AUVSI Unmanned Systems, Orlando, FL, USA, 12–15 May 2014. [Google Scholar]
  110. Fuentes-Peailillo, F.; Ortega-Farias, S.; Rivera, M.; Bardeen, M.; Moreno, M. Comparison of vegetation indices acquired from RGB and multispectral sensors placed on UAV. In Proceedings of the 2018 IEEE International Conference on Automation/XXIII Congress of the Chilean Association of Automatic Control (ICA-ACCA), Concepcion, Chile, 17–19 October 2018. [Google Scholar]
  111. Steven, M.D. The sensitivity of the OSAVI vegetation index to observational parameters. Remote Sens. Environ. 1998, 63, 49–60. [Google Scholar] [CrossRef]
  112. Guijarro, M.; Pajares, G.; Riomoros, I.; Herrera, P.J.; Burgos-Artizzu, X.P.; Ribeiro, A. Automatic segmentation of relevant textures in agricultural images. Comput. Electron. Agric. 2011, 75, 75–83. [Google Scholar] [CrossRef] [Green Version]
  113. Woebbecke, D.M.; Meyer, G.E.; Von Bargen, K.; Mortensen, D.A. Mortensen Color indices for weed identification under various soil, residue, and lighting conditions. Trans. ASAE 2013, 38, 259–269. [Google Scholar] [CrossRef]
  114. United States Naval Academy Point Cloud Data. Available online: https://www.usna.edu/Users/oceano/pguth/md_help/html/pt_clouds.htm (accessed on 7 October 2019).
  115. Potena, C.; Khanna, R.; Nieto, J.; Siegwart, R.; Nardi, D.; Pretto, A. AgriColMap: Aerial-ground collaborative 3D mapping for precision farming. IEEE Robot. Autom. Lett. 2019, 4, 1085–1092. [Google Scholar] [CrossRef]
  116. Dong, J.; Burnham, J.G.; Boots, B.; Rains, G.; Dellaert, F. 4D crop monitoring: Spatio-temporal reconstruction for agriculture. In Proceedings of the IEEE International Conference on Robotics and Automation, Singapore, 29 May–3 June 2017; pp. 3878–3885. [Google Scholar]
  117. Chebrolu, N.; Labe, T.; Stachniss, C. Robust long-term registration of UAV images of crop fields for precision agriculture. IEEE Robot. Autom. Lett. 2018, 3, 3097–3104. [Google Scholar] [CrossRef]
  118. Rusu, R.B.; Cousins, S. 3D is here: Point Cloud Library (PCL). In Proceedings of the IEEE International Conference on Robotics and Automation, Shanghai, China, 9–13 May 2011. [Google Scholar]
  119. Baugh, W.M.; Groeneveld, D.P. Empirical proof of the empirical line. Int. J. Remote Sens. 2008, 29, 665–672. [Google Scholar] [CrossRef]
  120. Kamilaris, A.; Prenafeta-Boldú, F.X. Deep learning in agriculture: A survey. Comput. Electron. Agric. 2018, 147, 70–90. [Google Scholar] [CrossRef] [Green Version]
  121. Sun, Y.; Xue, B.; Zhang, M.; Yen, G.G. Automatically designing CNN architectures using genetic algorithm for image classification. arXiv 2018, arXiv:1808.03818. [Google Scholar]
  122. Alsalam, B.H.Y.; Morton, K.; Campbell, D.; Gonzalez, F. Autonomous UAV with vision based on-board decision making for remote sensing and precision agriculture. In Proceedings of the IEEE Aerospace Conference, IEEE, Big Sky, MT, USA, 4–11 March 2017. [Google Scholar]
  123. Luxhøj, J.T. A socio-technical model for analyzing safety risk of unmanned aircraft systems (UAS): An application to precision agriculture. Procedia Manuf. 2015, 3, 928–935. [Google Scholar] [CrossRef]
  124. Jamoom, M.B.; Joerger, M.; Pervan, B. Sense and avoid for unmanned aircraft systems: Ensuring integrity and continuity for three dimensional intruder trajectories. In Proceedings of the 28th International Technical Meeting of the Satellite Division of The Institute of Navigation, Tampa, FL, USA, 14–18 September 2015. [Google Scholar]
  125. Sholes, E. Evolution of a UAV autonomy classification taxonomy. In Proceedings of the IEEE Aerospace Conference, Big Sky, MT, USA, 3–10 March 2007. [Google Scholar]
  126. FAA Recreational Flyers & Modeler Community-Based Organizations. Available online: https://www.faa.gov/uas/recreational_fliers/ (accessed on 7 October 2019).
  127. FAA Certificated Remote Pilots Including Commercial Operators. Available online: https://www.faa.gov/uas/commercial_operators/ (accessed on 7 October 2019).
  128. Petty, R.V. Drone use in aerial pesticide application faces outdated regulatory hurdles. Harvard J. Law Technol. Dig. 2018. Available online: https://jolt.law.harvard.edu/digest/drone-use-pesticide-application/ (accessed on 7 October 2019).
  129. Stoica, A.-A. Emerging legal issues regarding civilian drone usage. Challenges Knowl. Soc. 2018, 12, 692–699. [Google Scholar]
  130. Stöcker, C.; Bennett, R.; Nex, F.; Gerke, M.; Zevenbergen, J. Review of the current state of UAV regulations. Remote Sens. 2017, 9, 459. [Google Scholar] [CrossRef]
  131. UAV. Coach Master List of Drone Laws (Organized by State & Country). Available online: https://uavcoach.com/drone-laws/ (accessed on 7 October 2019).
  132. Helnarska, K.J.; Krawczyk, J.; Motrycz, G. Legal regulations of UAVs in Poland and France. Sci. J. Silesian Univ. Technol. Ser. Transp. 2018, 101, 89–97. [Google Scholar] [CrossRef]
  133. DJI Mavic 2 Pro. Available online: https://store.dji.com/product/mavic-2?vid=45291/ (accessed on 7 October 2019).
  134. Aqeel-Ur-Rehman; Abbasi, A.Z.; Islam, N.; Shaikh, Z.A. A review of wireless sensors and networks’ applications in agriculture. Comput. Stand. Interfaces 2014, 36, 263–270. [Google Scholar] [CrossRef]
  135. Wang, N.; Zhang, N.; Wang, M. Wireless sensors in agriculture and food industry—Recent development and future perspective. Comput. Electron. Agric. 2006, 50, 14. [Google Scholar] [CrossRef]
  136. Primicerio, J.; Matese, A.; Di Gennaro, S.F.; Albanese, L.; Guidoni, S.; Gay, P. Development of an integrated, low-cost and open-source system for precision viticulture: From UAV to WSN. In Proceedings of the EFITA-WCCA-CIGR Conference Sustainable Agriculture through ICT Innovation, Turin, Italy, 24–27 June 2013; pp. 24–27. [Google Scholar]
  137. Moribe, T.; Okada, H.; Kobayashl, K.; Katayama, M. Combination of a wireless sensor network and drone using infrared thermometers for smart agriculture. In Proceedings of the 15th IEEE Annual Consumer Communications and Networking Conference, Las Vegas, NV, USA, 12–15 January 2018. [Google Scholar]
  138. Uddin, M.A.; Mansour, A.; Le Jeune, D.; Ayaz, M.; Aggoune, E.H.M. UAV-assisted dynamic clustering of wireless sensor networks for crop health monitoring. Sensors 2018, 18, 555. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Example DJI Mavic 2 platform being used in a project at the Tennessee State University (TSU) Otis L. Floyd Nursery Research Center in McMinnville, TN involving automatic detection of biotic and abiotic stresses in nursery field productions. This unmanned aerial vehicle (UAV) platform will be equipped with a Raspberry Pi, 3D camera, and a multispectral camera to make up the unmanned aircraft system (UAS).
Figure 1. Example DJI Mavic 2 platform being used in a project at the Tennessee State University (TSU) Otis L. Floyd Nursery Research Center in McMinnville, TN involving automatic detection of biotic and abiotic stresses in nursery field productions. This unmanned aerial vehicle (UAV) platform will be equipped with a Raspberry Pi, 3D camera, and a multispectral camera to make up the unmanned aircraft system (UAS).
Agronomy 09 00618 g001
Figure 2. Original RGB and false color images of a crapemyrtle (Lagerstroemia spp.) field plot at the TSU Otis L. Floyd Nursery Research Center in McMinnville, TN captured with a UAS: (a) original RGB image; (b) excess red (ExR); (c) excess green (ExG); and (d) excess blue (ExB). Vegetation indices were visualized through a simple Python script written by the author.
Figure 2. Original RGB and false color images of a crapemyrtle (Lagerstroemia spp.) field plot at the TSU Otis L. Floyd Nursery Research Center in McMinnville, TN captured with a UAS: (a) original RGB image; (b) excess red (ExR); (c) excess green (ExG); and (d) excess blue (ExB). Vegetation indices were visualized through a simple Python script written by the author.
Agronomy 09 00618 g002

Share and Cite

MDPI and ACS Style

Hassler, S.C.; Baysal-Gurel, F. Unmanned Aircraft System (UAS) Technology and Applications in Agriculture. Agronomy 2019, 9, 618. https://doi.org/10.3390/agronomy9100618

AMA Style

Hassler SC, Baysal-Gurel F. Unmanned Aircraft System (UAS) Technology and Applications in Agriculture. Agronomy. 2019; 9(10):618. https://doi.org/10.3390/agronomy9100618

Chicago/Turabian Style

Hassler, Samuel C., and Fulya Baysal-Gurel. 2019. "Unmanned Aircraft System (UAS) Technology and Applications in Agriculture" Agronomy 9, no. 10: 618. https://doi.org/10.3390/agronomy9100618

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop