Next Article in Journal
An Accurate Calibration Method Based on Velocity in a Rotational Inertial Navigation System
Next Article in Special Issue
Optical Sensing of Weed Infestations at Harvest
Previous Article in Journal
AlGaN/GaN High Electron Mobility Transistor-Based Biosensor for the Detection of C-Reactive Protein
Previous Article in Special Issue
Spectral and Image Integrated Analysis of Hyperspectral Data for Waxy Corn Seed Variety Classification
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automated Mobile System for Accurate Outdoor Tree Crop Enumeration Using an Uncalibrated Camera

1
Department of Computer Science, University of California, Davis, CA 95616, USA
2
Department of Biological and Agricultural Engineering, University of California, Davis, CA 95616, USA
3
Department of Plant Sciences, University of California, Davis, CA 95616, USA
*
Authors to whom correspondence should be addressed.
Sensors 2015, 15(8), 18427-18442; https://doi.org/10.3390/s150818427
Submission received: 31 May 2015 / Revised: 6 July 2015 / Accepted: 20 July 2015 / Published: 28 July 2015
(This article belongs to the Special Issue Agriculture and Forestry: Sensors, Technologies and Procedures)

Abstract

:
This paper demonstrates an automated computer vision system for outdoor tree crop enumeration in a seedling nursery. The complete system incorporates both hardware components (including an embedded microcontroller, an odometry encoder, and an uncalibrated digital color camera) and software algorithms (including microcontroller algorithms and the proposed algorithm for tree crop enumeration) required to obtain robust performance in a natural outdoor environment. The enumeration system uses a three-step image analysis process based upon: (1) an orthographic plant projection method integrating a perspective transform with automatic parameter estimation; (2) a plant counting method based on projection histograms; and (3) a double-counting avoidance method based on a homography transform. Experimental results demonstrate the ability to count large numbers of plants automatically with no human effort. Results show that, for tree seedlings having a height up to 40 cm and a within-row tree spacing of approximately 10 cm, the algorithms successfully estimated the number of plants with an average accuracy of 95.2% for trees within a single image and 98% for counting of the whole plant population in a large sequence of images.

1. Introduction

To help cope with the rapid increase in the human population and future demands on worldwide food security, automation in agriculture is necessary. For example, there is a need to develop automatic systems for plant enumeration in fruit and nut tree seedling crops to save human resources and improve yield estimation. Most sensors used in agriculture have limited resolution, and cannot acquire the full scope of available plant and soil information. Advanced sensors, like cameras, that can characterize spatial and color information of natural objects play a crucial role in the future development of agricultural automation [1,2,3].
In the fruit and nut tree nursery industry, accurate counts of tree seedlings are very important in their production management and commerce [4,5]. Disease resistant tree rootstocks are planted from seeds in an outdoor nursery and they are later grafted to have fruit, which is a different cultivar from the disease resistant cultivar of the root system (i.e., to combine the best features of two cultivars). Variability in germination rate and consumption by birds create uncertainty in the number of marketable seedlings available for sale. Traditionally, human workers must manually count the seedlings each spring after they have grown large enough to be safe from the birds and the final crop stand is stable. This method is slow, tedious, and costly for workers to perform. Additionally, while this method can be accurate when carefully conducted, in practice, human error and bias are still present and can lower the accuracy of the final count, particularly when workers get fatigued or distracted.
Recently, methods for plant population and spacing measurement using machine vision have been introduced for different kinds of plants. A daylight sensing system is presented in [4] to measure early growth stage corn populations. The algorithms used in the system include steps for image sequencing to merge information between consecutive video frames, vegetation segmentation using a truncated ellipsoidal decision surface and a Bayesian classifier, and plant counting based on the total number of plant pixels and their median positions. The image sequencing step in this study does not consider the case of a camera perspective change, however. In [5], algorithms for automatically measuring corn plant spacing at early growth stages are introduced. Plant morphological features, plant color, and the crop row centerline are among multiple sources of information utilized by the algorithms for corn plant detection and localization. This work points out that the estimated interplant spacing errors are due to crop damage and sampling platform vibration, which caused mosaicking errors. A machine vision-based corn plant spacing and population measurement system is presented in [6]. Algorithm steps in this paper include image sequencing using SIFT (Scale Invariant Feature Transform) feature matching, vegetation segmentation based on color channels, corn plant center detection using a skeletonizing algorithm, and calculation of corn spacing and plant count. This algorithm yields satisfactory results with images captured from the top view. In [7], we had proposed a mobile platform that utilizes active optical sensing devices (LIDAR and light curtain transmission) to detect, localize, and classify trees. Promising results in our recent work helped system designers select the most reliable sensor for accurate detection of trees in a nursery and to automate labor-intensive tasks, such as weeding, without damaging crops.
In recent years, high-resolution remote-sensing techniques have been utilized in agricultural automation for counting mature trees. An automatic approach for counting olive trees is introduced in [8] with very high spatial remote sensing images. This approach contains two main steps: The olive trees are first isolated from other objects in the image based on a Gaussian process classifier with different morphological and textural features; candidate blobs of olive trees are then considered valid if their sizes are in a range specified by a prior knowledge of the real size of trees. In [9], a method for counting palm trees in Unmanned Aerial Vehicles (UAV) images is presented. To detect palm trees, SIFT keypoints are extracted from the images, and then analyzed by an Extreme Learning Machine (ELM) classifier. The ELM classifier uses prior knowledge trained on palm and non-palm keypoints. The output of this step is a list of palm keypoints that are then merged using an active contour method to produce the shape of each tree. Local binary patterns are used to distinguish palm trees from other objects based on their shapes. A general image processing method for counting any tree in high-resolution satellite images is described in [10]. Steps used in this method include HSI (hue, saturation, intensity) color space transform of the original image, image smoothing, thresholding of the extracted hue image, histogram equalization for HSI channels, candidate region detection, delineation, and tree counting.
This paper presents an uncalibrated camera system for fast and accurate plant counting in an outdoor field and incorporates a single high quality camera, an embedded microcontroller to automate the image capturing process, and a computer vision algorithm. The algorithm includes the steps of orthographic plant projection based on a perspective transform, plant segmentation using excessive green, plant detection by utilizing projection histograms, and plant counting that compensates for overlapping areas between consecutive images (to avoid double-counting). Both the camera and microcontroller were mounted on an ATV (all-terrain vehicle) and the images were analyzed offline.
Compared to previously described systems, the advantages of the proposed system are: It is easy to setup without requiring any camera calibration, it is robust to shadows in the background (e.g., soil, plant residue, and plants in other rows) and, among the target plants, it powerfully copes with noise and foliage occlusion, and is suited for use on a mobile vehicle in the field where row paths are rough and camera vibration common. The ATV platform was used in this paper to simulate the normal orchard tractors that the system was designed to be mount on. The objective was to have a system that can be mounted on a tractor and the counting done as part of the existing field operations. The most compatible existing operation is the fertilization step that is done at the same time as the counting. Figure 1 shows two small tractors performing the fertilization step at the same time as plant counting. This step is compatible because the tractors travel each row and are generating minimal dust. It is also the main argument against a single purpose UAV because by adding a machine vision module onto the front of the tractor shown in Figure 1, we can also count plants in an existing trip of a ground vehicle that is traveling each row.
This paper is organized as follows: In Section 2, we describe the system design and the proposed algorithm. The algorithm is detailed in Section 3 and Section 4 for plant counting for a single image and on an image sequence, respectively. Section 5 presents the experimental results of the system. Finally, this paper draws to a conclusion and discusses future work in Section 6 and Section 7, respectively.
Figure 1. Tractors are performing the fertilization step at the same time as the counting, at Sierra Gold Nurseries, CA, USA.
Figure 1. Tractors are performing the fertilization step at the same time as the counting, at Sierra Gold Nurseries, CA, USA.
Sensors 15 18427 g001

2. System Design and Algorithms

In the experimental system an ATV was used to simulate a tractor, as the base platform. The mobile vehicle had an attached arm to hold a camera, and a rotary shaft encoder was mounted on the wheel axle for odometry sensing. Image data are based on 24-bit digital color images taken by an electrically controlled, high-resolution, digital single-lens reflex camera (model EOS Rebel T3, Canon Inc., Tokyo, Japan). The camera was equipped with a zoom lens (model EF-S 18–55 mm 1:3.5–5.6 IS II, Canon Inc., Tokyo, Japan) aimed at the target plants and held fixed on the arm mounted to the mobile vehicle at an orientation of approximately 60° relative to the ground plane. An embedded microcontroller (model Raspberry Pi version 1 model B+, Raspberry Pi Foundation, UK) was used to activate the camera via a solid-state relay. The odometry signal was an input into the microcontroller to control the distance travelled between image acquisition events. A control algorithm was created to allow the microcontroller to trigger image acquisition by the camera at set spatial intervals. Because a difference in image resolution might dramatically affect the processing speed of the whole system from image transfer, plant segmentation, feature detection, and calculation of homography transformation matrix, the resolution 640 × 480 was selected as the best trade-off between speed and quality, so that accurate results can be obtained in an acceptable length of processing time. All camera parameters, including aperture, focal length, shutter speed, white balance, and ISO were manually set to have the best quality images in an outdoor scene. Figure 2 shows all details of the devices on the ATV, including the arm to mount the camera, the microcontroller with a relay and the wheel encoder.
The algorithm contains steps of perspective transform (with automatic determination of parameters) for orthographic plant projection, excessive green color segmentation, Gaussian smoothing, projection histogram, and local maxima detection for plant counting for a single image. To overcome potential problems with double-counting of plants in an image sequence, SURF (Speeded Up Robust Features) keypoint detection, SURF descriptor extraction [11], FLANN (Fast Library for Approximate Nearest Neighbors) descriptor matching [12], filtering of descriptor matches, and calculation of homography transform were utilized with GPU implementations. Figure 3 shows a block diagram of the proposed algorithm.
Figure 2. (a) The ATV with an arm to mount the camera, the wheel encoder, and the microcontroller to activate the camera via a relay; (b) the vehicle in operation capturing pictures of plants; and (c) manual counting of plants by a staff person.
Figure 2. (a) The ATV with an arm to mount the camera, the wheel encoder, and the microcontroller to activate the camera via a relay; (b) the vehicle in operation capturing pictures of plants; and (c) manual counting of plants by a staff person.
Sensors 15 18427 g002
Figure 3. Flowchart representation of the algorithm.
Figure 3. Flowchart representation of the algorithm.
Sensors 15 18427 g003

3. Plant Counting in a Single Image

3.1. Plant Straightening Using a Perspective Transform

Since the original image is collected using a perspective view, a correction method to straighten the plants (i.e., orthographic projection) for later processing in the steps of projection histogram calculation and local maxima detection is required. Existing studies [13,14] have investigated the affine rectification of the ground plane to make parallel world lines appear parallel in the rectified image. Only one vanishing point is needed in these methods when the camera is at a fixed angle tilting downward toward the ground. This affine rectification method has proven robust to create a virtual orthographic view of the scene. The method can be successfully applied to our system if a vanishing point could be found; however, parallel world landmarks do not exist in the field and thus requires another way of transforming the image. In this section, a perspective transform technique is presented using four predefined points in the source image space and four in the destination image space. The 3 × 3 perspective transform matrix P is solved so that:
[ x i y i 1 ] = P [ x i y i 1 ] = [ p 11 p 12 p 13 p 21 p 22 p 23 p 31 p 32 p 33 ] [ x i y i 1 ]
where i = {1, 2, 3, 4}; ( x i , y i ) and ( x i , y i ) are the i-th points in the source and destination images, respectively. The four points in the source image are defined as
( x 1 , y 1 ) = ( 0 , 0 )
( x 2 , y 2 ) = ( W 1 , 0 )
( x 3 , y 3 ) = ( W 1 , H 1 )
and  ( x 4 , y 4 ) = ( 0 , H 1 )
where W and H are the source image width and height, respectively. The four points in the destination image are calculated based on how the perspective changes with respect to x and y directions:
( x 1 , y 1 ) = ( δ x , δ y )
( x 2 , y 2 ) = ( W 1 δ x , δ y )
( x 3 , y 3 ) = ( W 1 + δ x , H 1 + δ y )
and  ( x 4 , y 4 ) = ( δ x , H 1 + δ x )
where δx and δy express how the perspective is needed to change (see Figure 4). These parameters can be determined automatically using our proposed algorithm that contains all steps from plant straightening to local maxima detection. The algorithm of auto-determination of perspective transform parameters is used only once on the first image of each dataset. The determined δx and δy are then applied to the remaining images in the set. This algorithm will be described in detail later in this section. Based on the obtained perspective transform matrix P, the destination image is transformed by
I p ( x , y ) = I ( p 11 x + p 11 y + p 13 p 31 x + p 32 y + p 33 , p 21 x + p 22 y + p 23 p 31 x + p 32 y + p 33 )
where Ip(x, y) and I(x, y) are the destination and source images, respectively. Figure 4 shows a sample image of ten small peach trees before and after performing a perspective transform. In this example, the average inclination angle of plants, with respect to x-axis, is significantly corrected from 74.81° to 82.77°.
Figure 4. An image before (a) and after (b) performing a perspective transform for plant straightening. The (red) solid lines show the plant inclination. The average inclination angle (with respect to x-axis) of the ten plants in the original image (a) is 74.81°; and it is 82.77° in the corrected one (b).
Figure 4. An image before (a) and after (b) performing a perspective transform for plant straightening. The (red) solid lines show the plant inclination. The average inclination angle (with respect to x-axis) of the ten plants in the original image (a) is 74.81°; and it is 82.77° in the corrected one (b).
Sensors 15 18427 g004

3.2. Plant Segmentation Based on Excessive Green

Different algorithms for soil, ground, and plant segmentation have been introduced in [15,16,17]. This step herein utilizes a fast and efficient method to segment green objects from the soil or background. The plants and the less-green background in the perspective transformed image were segmented by comparing the excessive green value [17] of each pixel to a certain threshold:
G ( x , y ) = { I p ( x , y ) ,     i f     ( 2 I p g ( x , y ) I p r ( x , y ) I p b ( x , y ) ) > T g 0 ,     o t h e r w i s e                                                                                            
where G(x, y) is the green-segmented image; I p g , I p r , and I p b are the green, red, and blue image channels of Ip(x, y), respectively; and Tg is a predefined threshold, which was fixed to 40 in this paper. This green color segmentation technique is very fast and effective in our system in segmenting plants from the backgrounds of soil, shadow, or non-green objects. Gaussian smoothing was applied to the thresholded excessive green image to eliminate background noise and reduce the effects of foliage occlusion. In this paper, based on the maximum plant height (approximately 40 cm) and the degree of foliage occlusion from plant to plant, the Gaussian kernel standard deviation size was fixed to 7. Figure 5 presents the green-segmented image of the transformed image in Figure 4 and the corresponding smoothed image.
Figure 5. A green-segmented image (a) and its Gaussian smoothed image (b) where plants are isolated.
Figure 5. A green-segmented image (a) and its Gaussian smoothed image (b) where plants are isolated.
Sensors 15 18427 g005

3.3. Projection Histogram and Local Maxima Detection

A vertical projection histogram [14] of the smoothed image is calculated in this section. It is constructed by projecting plant pixels along vertical lines (columns) in a region of interest (ROI). The number of bins in the histogram is equal to the number of columns in the image. In our system, because the camera is fixed on the mobile vehicle, we manually select the ROI to fit all plants in the field row. Let PV(Gg) be the vertical histogram of the smoothed image Gg, giving
PV(Gg) = {hj(Gg): j = 0, 1, …, W − 1}
Local maxima are detected by finding the histogram bins, where their gray values are greater than their neighbors as
L(i) = {hi(Gg): hi1(Gg) < hi(Gg) and hi+1(Gg) < hi(Gg), i = 1, 2, …, W − 2}
where L(i) is the gray value of the local maximum at bin i. The number of plants is then determined based on the number of local maxima. It is noted that the detection of the plants at the left and right borders explicitly followed Equation (7) and the definition of ROI, although objects outside the ROI occasionally appear to human eyes, as at the right bottom edges of Figure 5. The (red) numbers in Figure 5 represent the number of detected trees.

3.4. Auto-Determination of Perspective Transform Parameters

Notice that this step combines all steps from Section 3.1, Section 3.2 and Section 3.3 and it is done once on the first image of each dataset to have the optimal perspective transform parameters δx and δy for plant straightening and vertical histogram projection. The parameters δx and δy are estimated based on a list of detected local maxima as follows:
Δ ( δ x , δ y ) = arg max 0 δ x W / 4 0 δ y H / 2 k = 0 n L L ( k )
where nL is the number of local maxima and it is predefined as 10 for the image in Figure 5. This step particularly requires proper parameter tuning for the Gaussian filter to smooth the projection histogram and to have less noise in the detection of local maxima.

4. Plant Counting in an Image Sequence

Once the number of plants in each image is determined, the accumulated count in all images in a sequence is determined using a special method to avoid double-counting of trees between images. To serve this purpose, the homography transformation between every two successive images is considered. SURF is one of the best algorithms for keypoint detection and descriptor extraction and it is being used successfully in the applications of object recognition [18], image stitching [19], 3D reconstruction [20], and background motion compensation [21,22]. In this paper, SURF was utilized for the calculation of the homography transformation matrix, because it is fast to compute and has good performance when based on integral images and an integer approximation of the determinant of Hessian blob detector. The extracted SURF descriptors are matched between two consecutive images using FLANN, which allows fast and accurate nearest neighbor searches in high dimensional spaces. Furthermore, it can select the optimal matching parameters automatically without any tuning from users. Good matches are thenceforth obtained based on the condition of common distance, i.e., the matches having distances less than a minimum distance threshold will be discarded. In our case, a vehicle is moving along a field row and taking pictures of plants, two consecutive images of the same planar ground surface can be related by homography. The good matches obtained are used as the input to find a homography transformation between each set of successive images, without the need for camera calibration. Once the homography transformation matrix is found, camera translation with respect to vehicle’s horizontal movement can be extracted to estimate the overlap between successive images and avoid double-counting of plants. Defining It−1(x, y) = [xt−1, yt−1, 1]T and It(x, y) = [xt, yt, 1]T the image points as time t − 1 and t, a homography H is represented through
[ x t y t 1 ] = H [ x t 1 y t 1 1 ] = [ h 11 h 12 h 13 h 21 h 22 h 23 h 31 h 32 h 33 ] [ x t 1 y t 1 1 ]
where h13 is the coefficient of x-translation and that is the parameter we want to utilize for estimating the overlapping between successive images. It is noted that the scaling parameters h11 and h22 approach 1 because there are no zoom-in or zoom-out operators from the camera and the vehicle is travelling along the plant row. In our case, an affine homography is considered an appropriate model of image displacement and is a special type of a general homography where h31 and h32 are close to 0 and h33 approaches 1. In this paper, the Hessian threshold for the keypoint detector was fixed to 400, i.e., the features having a Hessian value larger than this threshold are retained. A RANSAC-based method was used to estimate the homography matrix so as to
H = arg min H i = 0 N [ ( x t i h 11 x t 1 i + h 12 y t 1 i + h 13 h 31 x t 1 i + h 32 y t 1 i + h 33 ) 2 + ( y t i h 21 x t 1 i + h 22 y t 1 i + h 23 h 31 x t 1 i + h 32 y t 1 i + h 33 ) 2 ]
where N is the total number of keypoints and ( x t i , y t i ) is the image point i at time t. Local maxima positions of plants are then compared to the x-translation parameter h13 to determine whether they are double-counted. Due to the effect of camera distortion (i.e., the plants at the image border are less straight than those closer to image center), this determination step requires a “buffer” for the plants at the image border to reduce the amount of counting errors. When h13 > 0, the comparison (ilb > h13) implies that the plant at the local maxima position il (obtained from the Equation (7)) is considered double-counted if its difference from the buffer b is greater than h13. When h13 ≤ 0, the comparison (il + b < W + h13) is to check if the plant at il is double-counted, where W is image width. The buffer b can be estimated automatically based on half of the average of distances between two neighbor local maxima.

5. Experimental Results

A computer (CPU model Core i7 at 3.4 GHz, Intel Co., Santa Clara, CA, USA, with 12-GB DDR3 random-access-memory) was used for all processing steps, except that an 1152 core GPU (model GeForce GTX 760, NVidia Co., Santa Clara, CA, USA) graphics card was utilized for implementing the GPU-based SURF descriptor extraction and FLANN matching algorithms. Experiments were conducted on 941 images containing 9915 juvenile Prunus persica L. “Nemaguard” peach trees (approximately 10.54 plants per image). Accounting for image overlap, the 941 images contained 2178 distinct peach trees when double-counting was eliminated. The images of the juvenile peach trees were collected from seven rows, with a 10 cm in-row plant distance between two neighboring plants, in an outdoor tree nursery (Sierra Gold Nurseries, Yuba City, CA, USA). Table 1 shows information for the datasets used in the experimental results, where Sets 1, 2, and 3 (namely, Group 1) were taken on the same day and Sets 4 to Set 7 (namely, Group 2) were acquired on another day. The maximum plant height was approximately 30 cm, and 40 cm in Groups 1 and 2, respectively. The overlap between two consecutive images was increased from approximately 60% in Group 1 to approximately 90% in Group 2. These plant images were selected as examples of plants covered by shadow (from human, other plants, or random objects), green and other objects in the background, small plants, and different plant sizes.
Figure 6 presents sample images showing examples of the green plant residue between the rows of plants, a second row of trees at the back (Figure 6a) and eight small plants (Figure 6b). In this study, the ground truth number of plants and overlap of images were measured manually. The number of small plants was defined based on a maximum plant height of 15 cm. It is noted that there were no plants imaged in shadow from Sets 3 to 7, and no additional objects on the background in Set 3. In the final design of our system on a tractor, a metal tunnel will be used to eliminate the current issues of shadows and green objects (i.e., the adjacent row) in the background. The purpose of presenting results of these critical issues here was to experimentally show the robustness of the proposed algorithm. In the experiments, all parameters were fixed (i.e., not optimized image by image) in order to have consistent results between different images and datasets.
Table 1. List of datasets and their detail information for experiments.
Table 1. List of datasets and their detail information for experiments.
Set 1Set 2Set 3Set 4Set 5Set 6Set 7Total
# imagesAll12112793150150150150941
Containing shadows1830000021
Containing green material on the background61046150150150150662
Containing other objects on the background35013136949
Containing small plants182112651134107
# plantsIn individual images14691481115413981402150515069915
In the image sequence609505525160136941492178
With shadows in individual images4670000053
Of small size in individual images264213651553160
Average image overlap59.2%64.7%53.8%89.4%91.1%93.8%90.4%
Figure 6. Sample images of green objects in the background (a) where there is plant residue on the soil and plants in the next row; and small plants (b).
Figure 6. Sample images of green objects in the background (a) where there is plant residue on the soil and plants in the next row; and small plants (b).
Sensors 15 18427 g006
Table 2 shows the average plant counting accuracies in both plant counting within a single image and for an image sequence, where double-counting was avoided. Within a single image, on average 0.51 counting mistakes out of 10.54 plants per image (i.e., an average accuracy of 95.2%) were observed. Accuracies obtained when excluding the challenging cases 95.4% (shadows), 95.4% (additional green objects), 95.3% (other objects), and 95.3% (small plants) were approximately equivalent to the overall performance, demonstrating the robustness of the method. Notice that the underlined numbers in Table 2 for the cases of “shadow” and “green objects” were taken from those without the exclusion. In an image sequence, when double-counting was minimized, an accuracy of 98% in all images was achieved. This accuracy is better than the single image performance because, on average, errors in a sequence of overlapped images compensate for one another. The plant counting error in every image is shown in Figure 7. The maximum estimation error was five plants and was mostly associated with high levels of foliage occlusion. By estimating the homography transformation between consecutive images, the proposed system also produced knowledge on the amount of image overlap. Comparing the actual overlap to that estimated by homography, we obtained an average error of only 2.54%. The average processing time for all steps in the software algorithm was approximately 300 ms per image pair.
Table 2. Average counting accuracy and the estimated image overlap using a homography transform.
Table 2. Average counting accuracy and the estimated image overlap using a homography transform.
Set 1Set 2Set 3Set 4Set 5Set 6Set 7Avg.*
Within single imagesw.r.t. individual images95.8%95.5%96.4%93.8%93.6%95.8%95.3%95.2%
Avg. count errors per image0.510.520.440.570.600.420.470.51
Std. Dev. of count errors per image0.820.670.670.740.830.590.630.72
Excluding the case of shadows96.2%95.8%96.4%93.8%93.6%95.8%95.3%95.4%
Excluding the case of green objects on the background95.7%95.5%96.7%----95.4%
Excluding the case of other objects on the background95.6%95.6%96.4%93.9%93.7%96.1%95.8%95.3%
Excluding the case of small plants95.8%95.4%96.5%94.0%93.9%96.2%95.4%95.3%
For an image sequence99.2%98.2%99.2%96.3%95.6%99%98.6%98%
Estimated image overlap57.1%62.1%58.7%87.6%88.6%91.7%88.6%
*: The final average values were calculated across all datasets.
Figure 7. Tree counting errors in 941 images.
Figure 7. Tree counting errors in 941 images.
Sensors 15 18427 g007
Comparisons in terms of counting accuracy and system characteristics between our method and the other two methods in [4,6] (where corn plants were used) are shown in Table 3. The results are presented with respect to counting in individual images, and for an image sequence using our peach tree data. In the method [4], the block matching based image sequencing algorithm was not applicable to our data, which were captured from a camera held at an angle of 60°. Due to a large number of incorrect matches yielded by the block matching based algorithm, our image sequencing algorithm (without the perspective transform) was substituted for the block matching in [4] to allow comparison of subsequent steps. The accuracies yielded by [4] were significantly lower than ours. The iterative rules based on the number of pixels and positions were sensitive to plant size and plant center locations. Additionally, these rules required parameter tuning for refining plant and background regions. Similarly, in the method from [6], when the top view was used, it led to plant counting errors when the direct image sequencing was utilized without the perspective transform. Using skeletonization for plant center detection, the method in [6] yielded a high error rate when there was foliage occlusion. For individual images, our system yielded an average accuracy of 95.2% (0.51 ± 0.72 count errors per image) compared to 86.9% (1.38 ± 1.34) for the method of [4] and 71.4% (2.05 ± 1.96) for the method in [6]. In the case of the image sequence, our total accuracy (98%) was substantially better than those of the other two methods (77.8% and 71.9%). It is worth mentioning that our method was able to work well with less than 60% image overlapping compared to 85% of [4].
Table 3. Accuracy and system characteristics comparison of the proposed method to [4,6].
Table 3. Accuracy and system characteristics comparison of the proposed method to [4,6].
The Method of [4]The Method of [6]The Proposed Method
Accuracy comparison (using peach tree seedling data)
Within individual imagesw.r.t. individual images86.9%71.4%95.2%
Avg. count errors per image1.382.050.51
Std. Dev. of count errors per image1.341.960.72
In an image sequence77.8%71.9%98%
System characteristics
Plant size (growth stage)V3 to V4 growth stages *Early to V3 growth stages *Early growth stage to 40 cm height
Camera viewTop viewTop viewPerspective view at an angle of 60°
Image overlap85%n/a54% to 91%
MethodImage sequencingBlock matching (substituted by our image sequencing method without perspective transform)SIFT feature matching, homography transformSURF descriptor extraction, RANSAC feature matching, homography transform
Plant segmentationBayesian classification on color spacesBayesian classification on color spacesExcessive green
Plant countingIterative rules based on the number of pixels and positionsSkeletonization for plant center detectionPerspective transform, Gaussian smoothing, projection histogram
*: The V3 growth stage in corn implies three leaves with visible leaf collars.

6. Conclusions

A digital imaging system for tree crop enumeration was successfully developed and tested. The system uses an embedded microcontroller mounted on an ATV (to simulate a tractor that will be used in actual practice the field) to receive the odometry signal and trigger the image acquisition. In this paper, the ability to automatically count a large number of small peach trees, with no human effort compared to manual counting, was demonstrated. The estimated count allows confirmation of the germination rate and final plant stand prior to budding in tree seedlings grown in outdoor nurseries. For the juvenile trees to a 40 cm height and 10 cm within-row plant spacing, the method successfully counted plants with an average accuracy of 95.2% in individual images and 98% for counting the whole plant population in a long image sequence. The method also provides robust performance for the cases of plants in shadow, green and other colored objects on the image background, small plants, and foliage occlusion.

7. Future Work

Future work should include (1) automating the system to perform real-time analysis and plant counting in the field; (2) improving and optimizing the algorithms to support larger seedlings; (3) counting multiple rows (Figure 8) simultaneously for a redundant check on the count or allowing a count with fewer passes through the field; (4) supporting measurements of stem width and plant height; and (5) plant phenotyping for tree crops in seedling nurseries and for mapping the location of each plant for individual plant care.
Figure 8. Sample image of rows for the purpose of counting multiple rows simultaneously.
Figure 8. Sample image of rows for the purpose of counting multiple rows simultaneously.
Sensors 15 18427 g008

Acknowledgments

Authors would like to thank Sierra Gold Nurseries, Cliff Beumel the Vice President of Sierra Gold Nurseries, and Rashpal Uppal their Production Manager for their assistance in conducting this research. We would also like to thank Burt Vannucci, Department of Biological and Agricultural Engineering, University of California, Davis, who helped build and test the system. This work was supported in part by the USDA National Institute of Food and Agriculture Hatch/Multistate project CA-D-BAE-7808-RR. Any opinions, findings, conclusions, or recommendations expressed in this publication are those of the author(s) and do not necessarily reflect the view of the National Institute of Food and Agriculture (NIFA) or the United States Department of Agriculture (USDA).

Author Contributions

David Slaughter and Bradley Hanson are the project leaders. Hanson was in charge of the field site coordination with Sierra Gold Nurseries and all aspects related to tree crop production and biology. Slaughter was in charge of the overall design of the mobile platform, computer vision hardware and general imaging processing approach and project management. The students, Andrew Barber, Amy Freitas, Daniel Robles, and Erin Whelan, were responsible for the detailed design of the platform. Burt Vannucci was system fabrication team leader and in charge of all equipment logistics, and the students performed the system fabrication. Erin and Daniel were in charge of vehicle operation. Andrew and Amy were in charge of the camera control system and the odometry sensor. Andrew was also the student team design leader. Slaughter, Vannucci, Barber, Freitas, Robles, and Whelan collected all the field data, and the students also made in-field ground truth counts. Thuy Tuong Nguyen was in charge of developing computer vision and image processing algorithms, writing the manuscript, and making image based ground truth counts. Sierra Gold Nurseries produced the crop and allowed researchers access to the tree nursery to test the system.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Pajares, G.; Peruzzi, A.; Gonzalez-de-Soto, P. Sensors in agriculture and forestry. Sensors 2013, 13, 12132–12139. [Google Scholar] [CrossRef] [PubMed]
  2. Emmi, L.; Gonzalez-de-Soto, M.; Pajares, G.; Gonzalez-de-Soto, P. Integrating sensory/actuation systems in agricultural vehicles. Sensors 2014, 14, 4014–4049. [Google Scholar] [CrossRef] [PubMed]
  3. Fernández, R.; Salinas, C.; Montes, H.; Sarria, J. Multisensory system for fruit harvesting robots. Experimental testing in natural scenarios and with different kinds of crops. Sensors 2014, 14, 23885–23904. [Google Scholar] [CrossRef] [PubMed]
  4. Shrestha, D.S.; Steward, B.L. Automatic corn plant population measurement using machine vision. Trans. ASAE 2003, 46, 559–565. [Google Scholar] [CrossRef]
  5. Tang, L.; Tian, L.F. Plant identification in mosaicked crop row images for automatic emerged corn plant spacing measurement. Trans. ASABE 2008, 51, 2181–2191. [Google Scholar] [CrossRef]
  6. Wang, C.; Guo, X.; Zhao, C. Detection of corn plant population and row spacing using computer vision. In Proceedings of the International Conference on Digital Manufacturing and Automation, Zhangjiajie, China, 5–7 August 2011; pp. 405–408.
  7. Garrido, M.; Perez-Ruiz, M.; Valero, C.; Gliever, C.J.; Hanson, B.D.; Slaughter, D.C. Active optical sensors for tree stem detection and classification in nurseries. Sensors 2014, 14, 10783–10803. [Google Scholar] [CrossRef] [PubMed]
  8. Bazi, Y.; Al-Sharari, H.; Melgani, F. An automatic method for counting olive trees in very high spatial remote sensing images. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium, Cape Town, South Africa, 12–17 July 2009; pp. II-125–II-128.
  9. Bazi, Y.; Malek, S.; Alajlan, N.; AlHichri, H. An automatic approach for palm tree counting in UAV images. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium, Quebec, QC, Canada, 13–18 July 2014; pp. 537–540.
  10. Kumar, D.G.; Padmaja, M. A novel image processing technique for counting the number of trees in a satellite image. Eur. J. Appl. Eng. Sci. Res. 2012, 1, 151–159. [Google Scholar]
  11. Bay, H.; Ess, A.; Tuytelaars, T.; van Gool, L. SURF: Speeded up Robust Features. Comput. Vis. Image Underst. 2008, 110, 346–359. [Google Scholar] [CrossRef]
  12. Muja, M.; Lowe, D.G. Scalable nearest neighbor algorithms for high dimensional data. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 2227–2240. [Google Scholar] [CrossRef]
  13. Sainz-Costa, N.; Ribeiro, A.; Burgos-Artizzu, X.P.; Guijarro, M.; Pajares, G. Mapping wide row crops with video sequences acquired from a tractor moving at treatment speed. Sensors 2011, 11, 7095–7109. [Google Scholar] [CrossRef] [PubMed]
  14. Nguyen, T.T.; Pham, X.D.; Song, J.H.; Jin, S.-H.; Kim, D.-K.; Jeon, J.W. Compensating background for noise due to camera vibration in uncalibrated-camera-based vehicle speed measurement systems. IEEE Trans. Veh. Technol. 2011, 60, 30–43. [Google Scholar] [CrossRef]
  15. Ribeiro, A.; Ranz, J.; Burgos-Artizzu, X.P.; Pajares, G.; Sanchez del Arco, M.J.; Navarrete, L. An image segmentation based on a genetic algorithm for determining soil coverage by crop residues. Sensors 2011, 11, 6480–6492. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Reina, G.; Milella, A. Towards autonomous agriculture: Automatic ground detection using trinocular stereovision. Sensors 2012, 12, 12405–12423. [Google Scholar] [CrossRef]
  17. Tian, L.; Slaughter, D.C.; Norris, R.F. Outdoor field machine vision identification of tomato seedlings for automated weed control. Trans. ASABE 1997, 40, 1761–1768. [Google Scholar]
  18. Ta, D.-N.; Chen, W.C.; Gelfand, N.; Pulli, K. SURFTrac: Efficient tracking and continuous object recognition using local feature descriptors. In Proceedings of the IEEE International Conference on Computer Vision Pattern Recognition, Miami, FL, USA, 20–25 June 2009.
  19. Juan, L.; Gwun, O. SURF applied in panorama image stitching. In Proceedings of the International Conference on Image Processing Theory Tools and Applications, Paris, France, 7–10 July 2010; pp. 459–499.
  20. Segundo, M.P.; Gomes, L.; Bellon, O.R.P.; Silva, L. Automating 3D reconstruction pipeline by SURF-based alignment. In Proceedings of the IEEE International Conference on Image Processing, Orlando, FL, USA, 30 September–3 October 2012; pp. 1761–1764.
  21. Nguyen, T.T.; Pham, X.D.; Jeon, J.W. Scaling-translation parameter estimation using Genetic Hough Transform for background compensation. KSII Trans. Internet Inf. Syst. 2011, 5, 1423–1443. [Google Scholar] [CrossRef]
  22. Chen, Y.; Zhang, R.H.; Shang, L. A novel method of object detection from a moving camera based on image matching and frame coupling. PLoS ONE 2014, 9, e109809. [Google Scholar] [CrossRef] [PubMed]

Share and Cite

MDPI and ACS Style

Nguyen, T.T.; Slaughter, D.C.; Hanson, B.D.; Barber, A.; Freitas, A.; Robles, D.; Whelan, E. Automated Mobile System for Accurate Outdoor Tree Crop Enumeration Using an Uncalibrated Camera. Sensors 2015, 15, 18427-18442. https://doi.org/10.3390/s150818427

AMA Style

Nguyen TT, Slaughter DC, Hanson BD, Barber A, Freitas A, Robles D, Whelan E. Automated Mobile System for Accurate Outdoor Tree Crop Enumeration Using an Uncalibrated Camera. Sensors. 2015; 15(8):18427-18442. https://doi.org/10.3390/s150818427

Chicago/Turabian Style

Nguyen, Thuy Tuong, David C. Slaughter, Bradley D. Hanson, Andrew Barber, Amy Freitas, Daniel Robles, and Erin Whelan. 2015. "Automated Mobile System for Accurate Outdoor Tree Crop Enumeration Using an Uncalibrated Camera" Sensors 15, no. 8: 18427-18442. https://doi.org/10.3390/s150818427

Article Metrics

Back to TopTop