Next Article in Journal
An Enhanced Joint Hilbert Embedding-Based Metric to Support Mocap Data Classification with Preserved Interpretability
Previous Article in Journal
Measuring Spatiotemporal Parameters on Treadmill Walking Using Wearable Inertial System
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Identifying the Branch of Kiwifruit Based on Unmanned Aerial Vehicle (UAV) Images Using Deep Learning Method

College of Mechanical and Electronic Engineering, Northwest Agriculture & Forestry University, Xi’an 712100, China
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(13), 4442; https://doi.org/10.3390/s21134442
Submission received: 10 June 2021 / Revised: 23 June 2021 / Accepted: 25 June 2021 / Published: 29 June 2021
(This article belongs to the Section Intelligent Sensors)

Abstract

:
It is important to obtain accurate information about kiwifruit vines to monitoring their physiological states and undertake precise orchard operations. However, because vines are small and cling to trellises, and have branches laying on the ground, numerous challenges exist in the acquisition of accurate data for kiwifruit vines. In this paper, a kiwifruit canopy distribution prediction model is proposed on the basis of low-altitude unmanned aerial vehicle (UAV) images and deep learning techniques. First, the location of the kiwifruit plants and vine distribution are extracted from high-precision images collected by UAV. The canopy gradient distribution maps with different noise reduction and distribution effects are generated by modifying the threshold and sampling size using the resampling normalization method. The results showed that the accuracies of the vine segmentation using PSPnet, support vector machine, and random forest classification were 71.2%, 85.8%, and 75.26%, respectively. However, the segmentation image obtained using depth semantic segmentation had a higher signal-to-noise ratio and was closer to the real situation. The average intersection over union of the deep semantic segmentation was more than or equal to 80% in distribution maps, whereas, in traditional machine learning, the average intersection was between 20% and 60%. This indicates the proposed model can quickly extract the vine distribution and plant position, and is thus able to perform dynamic monitoring of orchards to provide real-time operation guidance.

1. Introduction

In kiwifruit orchards, the acquisition of the plant growth status is particularly important for managers, and can help reduce the cost of orchard management and improve the efficiency of resource utilization. With the support of 3S (RS, GPS, GIS) technology, accurate remote sensing information of the orchard vegetation can be obtained, such as plant location and coverage [1]. Diversified information obtained via online methods enables managers to more accurately understand the growth of plants and control the spraying of pesticides [2], and make correct judgments regarding fertilization, pruning, and harvesting. Therefore, there is an urgent need to develop a fast, nondestructive, and stable technology to obtain plant information in kiwifruit orchards.
The unmanned aerial vehicle (UAV) is a new low-altitude remote sensing platform that can provide ultra-high-resolution images, flexibility for planning according to the weather and other factors, and dynamic remote sensing information [3]. Due to their high resolution and real-time characteristics, UAVs are a popular technology for obtaining large-scale plant information from UAV images. Different sensors that are carried by UAVs have been used to identify and count trees [4], and to determine the tree height and crown size [5,6]. This form of information collection, which is characterized by low data accuracy, is often used in the field of forestry management, which does not require accurate management of each tree. However, for the fine management of orchards, knowledge regarding the distribution of the plants is required, and the position and growth of each plant must be monitored, Thus, it is necessary to use more accurate sensors and more accurate plant phenotype processing algorithms. In recent years, the combination of low-altitude UAV remote sensing data, and deep learning data collection and processing, has significantly promoted the detailed management of orchards. For example, to count the number of citrus trees in an orchard [7], a simple CNN was designed and implemented. Based on the high-resolution images collected by the UAV, the CNN, which was a target recognition method, was able to separate citrus trees from other tree species. In orchard remote sensing, only simple parameters of trees, such as the crown size, can be obtained. To obtain more abundant information about the fruit trees from remote sensing images, Ampatzidis et al. extracted the number, crown size, and health index of fruit trees [8]. At present, the plant growth status obtained by UAV is often limited to the flourishing period of the plant. Thus, a valuable research area is the identification of changes in plants’ annual growth using UAV remote sensing, and the linking of these changes. In the plant growth model established in this study, UAV real-time images were employed to judge the growth status of plants at any time of the year. Subsequently, the labor intensity was significantly reduced, and errors arising from manual judgment could be avoided.
To date, few reports have been published regarding the extraction of branches. Wu et al. [9] proposed an extraction method for the crowns of bare trees. The method used the target recognition and segmentation of the CNN to realize the extraction and calculation size of a bare apple tree crown. Although this method was mainly focused on trees, the branches of apple trees are shorter and stronger, and artificial trellises are not used to fix them. Actinidia is a kind of liana that has a slender stem and needs to grow on scaffolds. Trellis planting is a common planting method. It is difficult to determine the plant position and canopy distribution in closed orchards when the fruit are maturing. Obtaining the distribution of the vines in advance is helpful for predicting the growth and guidance of the pruning. However, analyzing the vines of lianas is usually more challenging than analyzing the branches of trees. This is because: (1) the stems of the lianas are slender and irregular; (2) steel wires and slender stems are mixed among the scaffolds, which makes it difficult for human experts to accurately distinguish them; (3) and the ground is covered by grass, the background is not clear, and the texture is complex. The above reasons result in a diverse and misleading pattern for kiwifruit vine segmentation.
This study proposes an effective plant location distribution model (PLD-M) to interpret kiwifruit remote sensing images obtained using a low-altitude UAV. The three algorithms constitute the basis of the kiwifruit vine data acquisition and analysis model. First, according to the YOLOv3 deep learning target recognition network [10,11], the kiwifruit trunk in the UAV image is recognized and located. The improved version of this algorithm is also often used in the recognition of small targets in the agricultural field, such as pinecones [12], cherries [13], and apples [14]. The second key method, PSPnet [15], is a residual CNN used to extract the kiwifruit vines from the background using semantic segmentation. The third key approach, used to resample the region with vine pixels from the semantic segmentation results, provides the possible coverage and coverage gradient of the kiwifruit canopy. Next, the distribution of the kiwifruit vines and plants in the field is obtained by combining the location and gradient information. The specific objectives of this study include the following: (1) Compare and explore the advantages and potential of the kiwifruit field location and distribution models that are based on the low-altitude remote sensing images and deep learning modeling, in comparison to the support vector machine (SVM) and random forest (RF) classifiers. (2) To compare the effects of the different thresholds and sampling sizes on the PLD-M intersection over union (IoU). (3) To provide a fast and accurate method to obtain kiwifruit plant information in orchards. To verify the effectiveness of the PLD-M, a training set and a plant test set with 600 calibration images and 40 high-definition images, respectively, were constructed. We have published the dataset on GitHub: https://github.com/eletricsheep/PLD-M/tree/main. Last accessed at 23 June 2021.
To establish PLD-M model, the collected data were processed as follows:
(1)
The images were spliced by Pix4D to generate an orthophoto;
(2)
The orthophoto image was clipped to the size suitable for network training, and 300 images were selected to make the data set;
(3)
Three kinds of classifiers were trained by data set, and the trained classifiers were used to process the image to obtain the segmented image;
(4)
Different sampling size S and threshold level T were set to process the segmented image and count the IoU;
(5)
According to the change of IoU, the best parameters of S and T were selected to test 40 test images to verify the performance of the model;
(6)
Finally, the PLD-M model was formed.

2. Materials and Methods

2.1. Experimental Sites

The experimental site is located at the kiwifruit experimental station of Northwest Agricultural and Forestry University, Meixian County, Baoji City, Shaanxi Province, China (107°59′31.4443″ N, 34°7′27.8819″ E, elevation 643.22 m). The plant varieties in this experiment site were Heywood, with tree ages of 5–6 y, a plant spacing of 3 m, and a row spacing of 4 m. The area of the collection experiment site was 8814 m2 (Figure 1). To obtain a clear image of bare vines, the UAV images were collected at 14:32 on 14 January 2021, local time; the leaves of the plant had shed completely, and residual branches were left around the plant.
In this study, four rotor UAV (DJI phantom 4pro V2.0) was employed to aerial photography. Due to its short endurance and instability [16], both the course and side overlap were set at 80%. Table 1 shows the aerial photography and image parameters.
The UAV took 434 remote sensing images in total. The original orthophoto images were obtained by Pix4dmapper and mosaic with the CS_WGS_1984 coordinate system.

2.2. Brief Introduction of the PLD-M

The PLD-M can be divided into three phases, as shown in Figure 2. In the first stage, the position of the kiwifruit in the orchard was determined. During this stage, the most important task was to identify and calibrate the main position of the kiwifruit plants using target recognition. Therefore, we constructed a kiwifruit plant target detector using YOLOv3 (the network has an excellent small-scale target recognition performance). Then, we obtained the number of kiwifruit plants, the location of each tree, and the spatial distribution of each tree in the UAV image. The range of the vine pixels in the image was extracted. This study compared the performance of PSPnet, SVM, and RF to obtain the optimal algorithm as the segmentation model of the vines, and separate the vines from the background. Finally, the entire image was resampled and normalized based on the vine pixel data. The ratio of the vine pixels was returned to each sampling size, and spliced into the distribution map with gradient information and the location image that was obtained in the first stage. Finally, the location distribution of the output plants was determined. The PLD-M integrates the detection, segmentation, and distribution range prediction of the plants. Therefore, it is possible to extract more information from the UAV images, such as the number, location, and position of vines, and the possible canopy distribution range. The PLD-M is relatively independent in each stage, but each stage is necessary to complete the canopy prediction. Details of the three stages are given in Section 2.3 and Section 2.4.

2.3. Extraction of the Kiwifruit Vine

2.3.1. Random Forest Image Segmentation

The RF classifier, an ensemble-learning algorithm, was first proposed by Leo Breiman and Adele Cutler in 1995. The RF classifier contains multiple decision trees, and its output category is determined by the mode of the output category of the individual trees. As a result, it can generate a classifier with high accuracy for many kinds of input data. Because there is no complex parameter adjustment, it is widely used in remote sensing image classification, such as crop nitrogen content estimation [17] and land use classification [18].

2.3.2. Support Vector Machine Image Segmentation

SVM [19], a class of generalized linear classifiers, is a supervised learning method for the binary classification of data memory. The related theories and problems of SVM were first proposed in 1964, and a series of improved algorithms were rapidly developed in the 1990s. SVM is based on the principle of structural risk minimization, which has many advantages: it can avoid overlearning problems, has excellent generalization ability, is superior to other algorithms in terms of small samples and unbalanced data sets results [20], has a fast classification speed, and the results are easy to explain. This algorithm is also often used in the detection of agricultural diseases [21], pests [22], and other fields [23].

2.3.3. Deep Semantic Segmentation

The pyramid scene parching network (PSPnet) is an effective optimization strategy for deep Resnet [24] development, which monitors the change in the loss value. Because of the low position information of the high-level features in a deep network, it improves the content of the high-level features using a special fusion of the multi-scale features, and then uses a conditional random field (CRF) to process the segmentation results. PSPnet divides the full revolutionary network (FCN) images into four scales (1 × 1, 2 × 2, 3 × 3, 6 × 6) and then adds them to the input image to improve the prediction accuracy. The training and testing processes of the model are illustrated in Figure 3.
The calibration image of the dataset was labeled with the image after cutting 417 × 417 pixels. The spatial resolution of the image was 0.02 m, and the label file was made with Labelme. A total of 300 calibration images were amplified and input into the network for training. The experiment was performed on a Windows 10 professional operating system (Microsoft). This was achieved by using the pytorch1.7.1 framework that was built by Anaconda 3.0, and we built the network under Python 3.6.10 and Cuda11.1 to accelerate training. The details of the equipment are listed in Table 2.
To optimize and train the PSPnet deep semantic segmentation network, a training cycle of 100 generations was established and split into two stages, with a different Batch_size and Learning_rate. The specific training parameters are presented in Table 3.
We established two stages for the training of the model. The first stage aimed to improve the training speed, accelerate the convergence of the model; the learning rate was set as 10 4 and batch_size was eight. In the second stage, to identify the optimal solution, the learning rate was set to 10 5 and batch_size was four. For this study, training was conducted on VOC2007, which has two categories: background and kiwifruit vine. To reduce the gap between the predicted pixel class and the expected pixel class, we used the cross-entropy Function (1) to train the model:
L ( p , q ) = i n p i n log q i n
where p and q represent the expected pixel category and the predicted pixel category, respectively; i is the pixel; n is the category; p i n and q i n represent whether pixel i is classified in n (1 if classified into n, 0 if not). The classification accuracy index of the image was measured by the mean average precision (mAP). This represents the average value of all of the categories, as a percentage, i.e., the ratio of the number of correctly classified pixels of each type to all of the categories, as in Equation (2):
m A P = 1 n c i p i j p j
where n c represents the number of all the categories and p j represents the number of class i pixels that are predicted as class j.

2.4. Plant Location and Distribution

2.4.1. Resampling Processing

After the segmentation algorithm, the pixels containing the vines in images were demarcated; however, because the distribution of the vines is divergent from the center to the edge, only the discrete pixels of the vines can be obtained from the images. This results in a large amount of “pepper and salt noise” after classification.
To address the problems discussed above, firstly, the remote sensing image was binarized using three classification methods. Secondly, the binary clipping image was clipped according to the sampling size; the sampling size was represented by the letter S, and the proportion of vine pixels was counted. Finally, the resampled image was reconstructed by gradient filling. In this method, the interference pixels with large sparse distribution were screened out, and the influence on the distribution prediction was reduced. The resampling normalization process is shown in Figure 4.
The sampling size S is shown as:
S = ( 35 , 40 , 45 , 50 , 55 , 60 , 65 , 70 , 75 , 80 , 85 , 90 )
During resampling normalization, we can take the number of pixels in a single channel image as the denominator of the normalization, and the ratio of the normalized pixels is expressed by the letter R, as shown in Equation (4):
R = p i p i + p j
where p i   and   p j represent the ratio of pixels that are divided into kiwifruit vines and the ground class, respectively, which reflects the proportion of the vine to the cut image.

2.4.2. Coordinate and Resampling

Using the trained YOLOv3 network, the main stem of the kiwifruit was identified and selected from the image. YOLOv3 is a target recognition network based on Darknet_53; it is a full convolutional network (FCN) with 53 convolution layers. Compared with other target recognition networks, the YOLOv3 network has three scale outputs (13 × 13, 26 × 26, 52 × 52) in the down-sampling process, which makes it more effective for different sizes of targets, and particularly small targets. To distinguish the bare land area and the vine area more clearly, the gradient map palette uses the Seaborn Python data visualization library. After overlapping and comparing the images, the vine density was divided into four levels, as shown in Table 4.
The YOLOv3 network outputs the image with the kiwifruit plant position (the plant position is the red box calibration; to enhance the display, this study conducted the secondary annotation through the red dot), and then fuses the image after semantic segmentation and resampling. The fusion process for the location and distribution images is shown in Figure 5.

3. Results

3.1. Analysis of the Image Processing Results

3.1.1. Accuracy Evaluation of the RF, SVM, and PSPnet

After image segmentation, we obtained the segmentation images of the RF, SVM, and deep semantic segmentation. According to the calculation, we obtained the mAP and signal-to-noise ratio (SNR) of each classification method, as shown in Table 5.
By comparing the mAP of the three segmentation methods, it can be observed that the pixel accuracy of SVM was approximately 85%. The accuracy of PSPnet was 71%, i.e., 14 percentage points lower than that of SVM. From the perspective of this data, the mAP of the two traditional machine learning segmentation methods is higher than that of deep semantic segmentation. An image of the classification result is shown in Figure 6.
The comparison of the segmentation results in Figure 6 shows that the noise for traditional machine learning is greater. The RF and SVM are more sensitive to the fine features than the depth semantic segmentation, and the distribution range of the vine pixels in the correct region is more accurate. However, in general, the quality of the segmentation image in traditional machine learning is not as high as that obtained by deep semantic segmentation. There are three reasons for this finding. First, the pixels obtained from machine learning are separated from each other, which results in a mixture of pixels in the background, the object, and a rough outline of the object edge. In addition, there is more noise at the edges. Second, because of the pruning that takes place in winter, there are more residual branches on the ground and more complex ground objects. The texture of the UAV image obtained in practice is more complex. The RF and SVM incorporate pixels that are similar to the label into the target object, which further increases the noise of the segmentation results. Finally, various objects (e.g., well covers, wires, and scaffolding wires) and the objects on the edge of the orchard are similar to the target objects. As demonstrated in Figure 7a, in comparison to machine learning, due to the integrity of the dataset annotation, deep semantic segmentation does not segment the ground stumps as objects. Furthermore, the segmented objects have a continuous pixel distribution, smooth contour, and no noise. They also do not include the objects at the edge of the orchard, such as well covers and scaffolding, in the objects, as shown in Figure 7c.

3.1.2. Advantages of Deep Semantic Segmentation

The average pixel accuracy of PSPnet is lower than that of the former two traditional machine learning segmentation methods. However, in the actual classification results, the results that are obtained by deep semantic segmentation have a better continuity than in traditional machine learning, and the recognition of objects is also better than that of traditional machine learning. In contrast to previous methods of remote sensing deep learning (such as field [25], citrus tree [26,27], and orchard remote sensing), kiwifruit vine recognition presents unique characteristics. First, because of the high density and uniform crops that are planted in the field, the target objects need to be recognized in the field crop images, which generally have continuous and uniform features and an obvious texture. In the process of kiwifruit vine recognition, the background pixels generally have an obvious texture, but the target objects are extremely small and difficult to distinguish. Second, compared with the larger and obvious branches of large fruit trees, kiwifruit trees are smaller and irregular.
The training of deep learning is robust. Due to the particularity of the features of kiwifruit orchards, PSPnet mainly avoids the misclassification of pixels. In the process of compiling the dataset, the range of vine pixels is relatively small, which makes it difficult to label, resulting in different error pixels in the labeled image. This small number of error pixels results in significant changes in the segmentation accuracy of traditional machine learning. However, when training in the PSPnet network, although it is difficult to label the vine pixels, these wrong pixels do not significantly affect the segmentation accuracy. The training time of the deep semantic segmentation network continues to increase with the increase in the sample size, and is relatively stable.

3.2. Influence of Threshold Parameters on the Accuracy of the Distributed Images

To evaluate the image distribution accuracy, due to the existence of the resampling size we can use the Otsu iterative thresholding method and other methods to select the segmentation threshold for a single time. In this study, the binary segmentation of each sample size S was performed for eight scales. A total of 288 binary images were obtained. The threshold value ranged from 0 to 1, and the eight gradients were equally divided, as indicated by the letter T. The corresponding relationship between the threshold and threshold levels is shown in Table 6.
The distribution range of the real vines was compared to the continuous vines. The actual distribution range of the vines and the distribution range of the vines that were obtained after resampling were combined to calculate the IoU. The formula for the IoU is as follows:
I O U = P Q P Q
where P is the pixel range of the true vine distribution, and Q is the pixel range of the true vine distribution. The distribution map of the vines was binarized in the process of the IoU calculation, as shown in Figure 8. This is the binary image of the various classification methods under the S = 70 sampling size, and the T = 3 and T = 7 threshold levels.
Figure 8 shows the binary images of the three segmentation methods under the threshold levels of S = 70, and T = 3 and 7. The changes caused by the three different classification methods under the same parameter show that the pixel range of PSPnet segmentation is more concentrated, followed by that of SVM segmentation. As shown in Figure 8a,b,e, the pixels obtained by these two methods are more concentrated and show an island shape; the worst effect is that of RF segmentation. At a lower threshold level, the pixels are more discrete, such as in Figure 8c, but with the increase in threshold level, the useful information has been obscured by noise; Figure 8d. At the threshold level of T = 7, PSPnet can still maintain a relatively independent island distribution, whereas the SVM segmentation method results in increased noise. The reason for the above results is that, with the increase in the threshold level (that is, with the decrease in the threshold value), more pixels are allowed to be collected, which not only increases the number of useful pixels, but also increases the noise. Because the image SNR produced by the PSPnet segmentation method is better than that of the other two machine learning methods, due to the improvement in the threshold level, RF and SVM add a large amount of noise, whereas the pixel change of PSPnet is relatively stable.
In conclusion, the distribution of the pixels in the binary image shows that, with an increase in the threshold level (and thus a decrease in the threshold value), the binary image contains more information. The area of the yellow region (vine distribution region and noise region) changes with the threshold value. A higher threshold value can obtain continuous and low-noise binary segmentation results, but may also cause the original correctly predicted segmentation. From this, the layout area is removed, and the IoU of each binary image is calculated under eight threshold levels. The trend of the IoU variation with the accuracy is the same for the different sizes. Only the variation curves of the IoU with the threshold level T under the sampling sizes S = 35 and S = 70 are listed in the text, as shown in Figure 9. (Detailed data are shown in Table A1, Table A2, Table A3, Table A4, Table A5, Table A6, Table A7, Table A8, Table A9, Table A10, Table A11 and Table A12. Only the data in A1 and A9 are shown here).
The experimental data show that regardless of the resampling parameters, the order of the IoU that is obtained by the three classification methods from large to small is PSPnet, SVM, and RF classification. When the threshold level T = 5, the maximum IoU is obtained using PSPnet. At a small threshold level, the difference in the IoU that is obtained by the three segmentation methods is not obvious. However, with an increase in the threshold level, the IoU curves that are obtained by the three segmentation methods exhibit two trends. The IoU curve of the deep semantic segmentation shows a growing trend and remains stable after reaching the maximum IoU, and changes little with the threshold. The change curves of the two traditional machine learning methods show a monotonic decreasing trend, and the accuracy of the RF classification declines more rapidly.
There are two reasons for the above results. First, because the result of the depth network segmentation has continuous pixels and less noise distribution, the gradient of the pixels in the different regions of the resampled image is different. Even at a higher threshold level, it is still easy to extract different pixels in the resampled image. Second, traditional machine learning has some disadvantages in distinguishing complex textures and similar objects. Even through median filtering, it still cannot obtain good results (vine pixels and noise are similar in size, smaller filter cores cannot remove obvious noise, and larger filter cores will lead to the loss of vine pixel information). Therefore, when the threshold level increases, the mixed background noise is also included, which leads to an enlargement of the union area and a decrease in the IoU.

3.3. Model Evaluation

Due to the comparison of the first sections, the PSPnet deep semantic segmentation network has obvious advantages over traditional machine learning in image quality and the IoU results of resampling. With sampling size S and threshold level T as independent variables, and IoU as the dependent variable, the IoU change surfaces of three segmentation methods are shown in Figure 10. (Detailed data are shown in Table A13, Table A14 and Table A15 of Appendix A).
Figure 10 shows the IoU changes of the three segmentation methods under different parameters. From Figure 10a, regardless of the sampling size S, the threshold level T = 5 can achieve better IoU. By selecting the threshold level T = 5, the performance of the IoU of PSPnet on 40 test sets after processing with different sampling sizes is shown in Figure 11 (each color curve represents a different test chart).
Figure 11 shows that, when the threshold level T = 5, the sampling parameters used to obtain a high IoU are mainly concentrated around the sampling size S = 70. Figure 12 shows the change curve of the true value and the predicted value of the 40 pictures in the test set when the sampling size is S = 70 and T = 5.

4. Discussion

4.1. Advantages and Limitations of the Model

In this study, an effective model (PLD-M) for kiwifruit orchard data acquisition from low-altitude UAV remote sensing images is proposed. This includes the detection and segmentation of kiwifruit vines, and the prediction of the possible canopy distribution. The use of a deep learning method increased the quality of the segmentation image, and intersection and union ratio, compared to those of traditional machine learning. The existing methods of dense tree canopy recognition and extraction can only obtain the number of fruit trees, canopy parameters, and other information. In contrast, the PLD-M model can not only count and locate kiwifruit plants, but also extract the vine distribution and possible canopy distribution of each plant. Compared with the apple branch extraction model proposed by J Wu et al. (2020), PLD-M can not only obtain the specific distribution of vines and the position of plants, but also extract the distribution density gradient of liana plants. The distribution range of the canopy is more refined, and the hollow area of the canopy caused by the external polygon of branches is avoided.
Due to the small size of the ground stump, camera lens distortion, and pan tilt stability, small distortions appear in the stitching process, which may cause tiny inconsistencies in the texture of different regions. In subsequent research, these problems could be addressed by improving the resolution of the remote sensing camera, using a lens with less distortion, and enhancing the pan tilt stability.
In this paper, we do not discuss the gap between the proposed PSPnet and other deep learning networks, such as that of Lucas Prado Osco et al. [28]. This is because, compared with the recognition of large areas, the mAP of deep learning network is not superior to traditional machine learning in the classification of kiwifruit vines. However, this study can optimize the data set and network input through feature fusion [29] and image fusion [30], and thus allow comparison of the differences in different neural networks in multi-feature and multi-channel images.

4.2. Research Significance and Prospect

After artificial pruning of kiwifruit plants, the originally dense vines show an uneven and sparse distribution. Although the sparse degree of vine distribution of a single plant is within the reasonable range of artificial control, the vine distribution among kiwifruit plants is uneven. This may lead to the overlap of the canopy between kiwifruit plants, thus affecting plant photosynthesis and, ultimately, yield. The orchard information is collected using a low-altitude UAV and input into PLD-M to obtain the distribution of vines of each plant and the distribution of vines among plants, thus allowing evaluation of the rationality of winter pruning and assisting in decision making. This can not only be used for the evaluation of pruning, but also for variable rate spraying and variable rate fertilization.
Possible future directions exist for this research. First, the model could be applied to the embedded system of UAVs for real-time orchard condition analysis, and thus to precision agriculture. Second, because the training of deep learning networks depends on Big Data, datasets should be further expanded and enriched in further study of the model. Third, for variable applications, models should be designed that can provide a greater amount of information, such as the elevation and overlap. Fourth, at present, the model is only suitable for kiwifruit plants, and future research can extend the model to other plants. Finally, the image information that is obtained by a visible light sensor is limited. To expand the amount of information that is collected, multispectral data will be introduced in future investigations.

5. Conclusions

In this study, an effective model (PLD-M) for kiwifruit orchard data acquisition from low-altitude UAV remote sensing images is proposed. This includes the detection and segmentation of kiwifruit vines, and prediction of the possible canopy distribution. The model is based on the YOLOv3 target recognition network and the PSPnet deep semantic segmentation network. The former extracts the trunk position of the kiwifruit plants and provides the location information of each plant, whereas the latter extracts the kiwifruit vines from the background. Then, the segmented image is input into a model to predict the canopy range using a resampling operation. Finally, the kiwifruit plant information, such as the vine position and canopy distribution, can be obtained from the low-altitude UAV images. To train and optimize the proposed model, this study constructed a training set that contained 300 labeled images to optimize and train the YOLOv3 network. We also constructed a training set that contained 300 labeled images to train and optimize the PSPnet deep semantic segmentation network that was used to extract the kiwifruit vines. To verify and optimize the accuracy of the canopy prediction that was obtained by PSPnet and resampling, another 40 high-definition images were selected from the UAV remote sensing images for manual truth labeling. The results showed that the canopy data that were predicted by the model were close to the artificial data (the IoU of the canopy distribution was between 70% and 95%). Therefore, the proposed model can accurately extract the vine distribution and plant position of each plant in a kiwifruit orchard, based on UAV data, allowing rapid, non-destructive, and stable monitoring of kiwifruit plants.

Author Contributions

Software, J.D.; validation, J.D.; investigation, X.Z., J.Z., S.P. and H.M.; data curation, J.D.; writing—original draft preparation, Z.N.; writing—review and editing, Z.N. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Key Research and Development Plan of China, grant number 2019YFD1002401.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. This data can be found here: [https://github.com/eletricsheep/PLD-M/tree/main].

Acknowledgments

This research was supported by National Key Research and Development Plan of China with Research Grant 2019YFD1002401.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

Appendix A

Table A1. The IoU of sample size S = 35 varies with the threshold level T.
Table A1. The IoU of sample size S = 35 varies with the threshold level T.
MethodsT = 1T = 2T = 3T = 4T = 5T = 6T = 7T = 8
PSPnet0.49690.61920.63800.63940.62920.62710.61970.6126
RF0.47420.46420.35590.25310.20840.18770.17640.1704
SVM0.47430.46040.44080.41190.38000.34940.32430.3056
Table A2. The IoU of sample size S = 40 varies with the threshold level T.
Table A2. The IoU of sample size S = 40 varies with the threshold level T.
MethodsT = 1T = 2T = 3T = 4T = 5T = 6T = 7T = 8
PSPnet0.48240.61520.64400.64470.63110.62790.62000.6172
RF0.46580.46300.35600.24630.20110.18280.17200.1660
SVM0.46580.45610.43440.41320.37520.34710.32030.3006
Table A3. The IoU of sample size S = 45 varies with the threshold level T.
Table A3. The IoU of sample size S = 45 varies with the threshold level T.
MethodsT = 1T = 2T = 3T = 4T = 5T = 6T = 7T = 8
PSPnet0.52830.63410.66570.67050.66400.65170.64430.6370
RF0.51360.51210.38090.25620.20930.18960.17920.1732
SVM0.51320.49920.47800.44210.39880.36110.33620.3155
Table A4. The IoU of sample size S = 50 varies with the threshold level T.
Table A4. The IoU of sample size S = 50 varies with the threshold level T.
MethodsT = 1T = 2T = 3T = 4T = 5T = 6T = 7T = 8
PSPnet0.52250.62870.67500.68080.67510.66290.65450.6435
RF0.51350.50520.38820.25700.20690.18830.17750.1717
SVM0.51290.50110.48100.45370.40960.36920.33170.3135
Table A5. The IoU of sample size S = 55 varies with the threshold level T.
Table A5. The IoU of sample size S = 55 varies with the threshold level T.
MethodsT = 1T = 2T = 3T = 4T = 5T = 6T = 7T = 8
PSPnet0.52300.64020.69140.71580.70610.68460.67610.6658
RF0.51300.51500.38310.25280.20640.18820.17880.1755
SVM0.51250.50140.48560.45650.41500.37890.34830.3310
Table A6. The IoU of sample size S = 60 varies with the threshold level T.
Table A6. The IoU of sample size S = 60 varies with the threshold level T.
MethodsT = 1T = 2T = 3T = 4T = 5T = 6T = 7T = 8
PSPnet0.51190.63180.68010.71280.69460.67560.66160.6550
RF0.50510.50400.39440.25300.20000.18140.17220.1685
SVM0.50440.49200.47730.43820.40120.36120.33560.3123
Table A7. The IoU of sample size S = 65 varies with the threshold level T.
Table A7. The IoU of sample size S = 65 varies with the threshold level T.
MethodsT = 1T = 2T = 3T = 4T = 5T = 6T = 7T = 8
PSPnet0.46600.59050.64340.66550.67130.64920.63060.6229
RF0.45810.45950.37280.23960.19230.17600.16810.1630
SVM0.45810.45410.42680.42370.39920.35970.33310.3074
Table A8. The IoU of sample size S = 70 varies with the threshold level T.
Table A8. The IoU of sample size S = 70 varies with the threshold level T.
MethodsT = 1T = 2T = 3T = 4T = 5T = 6T = 7T = 8
PSPnet0.49590.62690.71340.73130.75060.73050.70390.6844
RF0.48910.49860.38680.24740.19430.17720.16970.1651
SVM0.48910.47950.46560.45070.41710.37730.33890.3139
Table A9. The IoU of sample size S = 75 varies with the threshold level T.
Table A9. The IoU of sample size S = 75 varies with the threshold level T.
MethodsT = 1T = 2T = 3T = 4T = 5T = 6T = 7T = 8
PSPnet0.50630.62280.68880.72930.74320.72210.69610.6732
RF0.50420.51480.39680.25120.19540.17990.17190.1679
SVM0.50420.49410.47560.44910.40730.35750.32560.3050
Table A10. The IoU of sample size S = 80 varies with the threshold level T.
Table A10. The IoU of sample size S = 80 varies with the threshold level T.
MethodsT = 1T = 2T = 3T = 4T = 5T = 6T = 7T = 8
PSPnet0.52210.63110.69260.72300.72480.71670.69310.6863
RF0.51730.52590.39820.25390.20320.18330.17740.1724
SVM0.51620.50000.48440.44010.38700.34820.31870.2949
Table A11. The IoU of sample size S = 85 varies with the threshold level T.
Table A11. The IoU of sample size S = 85 varies with the threshold level T.
MethodsT = 1T = 2T = 3T = 4T = 5T = 6T = 7T = 8
PSPnet0.51770.61800.67220.69810.69380.68460.66740.6590
RF0.51500.51340.40090.25360.20240.18430.17710.1730
SVM0.51360.50930.49460.45110.40840.37640.34440.3192
Table A12. The IoU of sample size S = 90 varies with the threshold level T.
Table A12. The IoU of sample size S = 90 varies with the threshold level T.
MethodsT = 1T = 2T = 3T = 4T = 5T = 6T = 7T = 8
PSPnet0.50520.62040.67980.71530.73090.69980.69290.6794
RF0.50220.50910.34090.22660.18570.17510.16810.1650
SVM0.50220.49580.47350.44200.38750.33990.31120.2901
Table A13. The IoU of PSPnet varies with sampling size S and threshold level T.
Table A13. The IoU of PSPnet varies with sampling size S and threshold level T.
S = 35S = 40S = 45S = 50S = 55S = 60
T = 10.49690.48240.52830.52250.52300.5119
T = 20.61920.61520.63410.62870.64020.6318
T = 30.63800.64400.66570.67500.69140.6801
T = 40.63940.64470.67050.68080.71580.7128
T = 50.62920.63110.66400.67510.70610.6946
T = 60.62710.62790.65170.66290.68460.6756
T = 70.61970.62000.64430.65450.67610.6616
T = 80.61260.61720.63700.64350.66580.6550
S = 65S = 70S = 75S = 80S = 85S = 90
T = 10.46600.49590.50630.52210.51770.5052
T = 20.59050.62690.62280.63110.61800.6204
T = 30.64340.71340.68880.69260.67220.6798
T = 40.66550.73130.72930.72300.69810.7153
T = 50.67130.75060.74320.72480.69380.7309
T = 60.64920.73050.72210.71670.68460.6998
T = 70.63060.70390.69610.69310.66740.6929
T = 80.62290.68440.67320.68630.65900.6794
Table A14. The IoU of RF varies with sampling size S and threshold level T.
Table A14. The IoU of RF varies with sampling size S and threshold level T.
S = 35S = 40S = 45S = 50S = 55S = 60
T = 10.47420.46580.51360.51350.51300.5051
T = 20.46420.46300.51210.50520.51500.5040
T = 30.35590.35600.38090.38820.38310.3944
T = 40.25310.24630.25620.25700.25280.2530
T = 50.20840.20110.20930.20690.20640.2000
T = 60.18770.18280.18960.18830.18820.1814
T = 70.17640.17200.17920.17750.17880.1722
T = 80.17040.16600.17320.17170.17550.1685
S = 65S = 70S = 75S = 80S = 85S = 90
T = 10.45810.48910.48910.51730.51500.5022
T = 20.45950.49860.49860.52590.51340.5091
T = 30.37280.38680.38680.39820.40090.3409
T = 40.23960.24740.24740.25390.25360.2266
T = 50.19230.19430.19430.20320.20240.1857
T = 60.17600.17720.17720.18330.18430.1751
T = 70.16810.16970.16970.17740.17710.1681
T = 80.16300.16510.16510.17240.17300.1650
Table A15. The IoU of SVM varies with sampling size S and threshold level T.
Table A15. The IoU of SVM varies with sampling size S and threshold level T.
S = 35S = 40S = 45S = 50S = 55S = 60
T = 10.47430.46580.51320.51290.51250.5044
T = 20.46040.45610.49920.50110.50140.4920
T = 30.44080.43440.47800.48100.48560.4773
T = 40.41190.41320.44210.45370.45650.4382
T = 50.38000.37520.39880.40960.41500.4012
T = 60.34940.34710.36110.36920.37890.3612
T = 70.32430.32030.33620.33170.34830.3356
T = 80.30560.30060.31550.31350.33100.3123
S = 65S = 70S = 75S = 80S = 85S = 90
T = 10.45810.48910.50420.51620.51360.5022
T = 20.45410.47950.49410.50000.50930.4958
T = 30.42680.46560.47560.48440.49460.4735
T = 40.42370.45070.44910.44010.45110.4420
T = 50.39920.41710.40730.38700.40840.3875
T = 60.35970.37730.35750.34820.37640.3399
T = 70.33310.33890.32560.31870.34440.3112
T = 80.30740.31390.30500.29490.31920.2901

References

  1. Guo, X. Application and development prospect of 3S technology in precision agriculture. Agric. Technol. 2020, 40, 41–43. [Google Scholar]
  2. Shahzad, A.; Arfeen, Z.A.; Khan, M.A.; Ahmed, T.K.; Tauqeer, A. SLM-OJ: Surrogate Learning Mechanism during Outbreak Juncture. Int. J. Mod. Trends Sci. Technol. 2020, 6, 162–167. [Google Scholar]
  3. Varela, S.; Dhodda, P.R.; Hsu, W.H.; Prasad, P.V.V.; Assefa, Y.; Peralta, N.R.; Griffin, T.; Sharda, A.; Ferguson, A.; Ciampitti, I.A. Early-Season Stand Count Determination in Corn via Integration of Imagery from Unmanned Aerial Systems (UAS) and Supervised Learning Techniques. Remote Sens. 2018, 10, 343. [Google Scholar] [CrossRef] [Green Version]
  4. Goldbergs, G.; Maier, S.W.; Levick, S.R.; Edwards, A. Efficiency of Individual Tree Detection Approaches Based on Light-Weight and Low-Cost UAS Imagery in Australian Savannas. Remote Sens. 2018, 10, 161. [Google Scholar] [CrossRef] [Green Version]
  5. Mu, Y.; Fujii, Y.; Takata, D.; Zheng, B.Y.; Noshita, K.; Honda, K.; Ninomiya, S.; Guo, W. Characterization of peach tree crown by using high-resolution images from anunmanned aerial vehicle. Hortic. Res. 2018, 5, 74. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Panagiotidis, D.; Abdollahnejad, A.; Surovy, P.; Chiteculo, V. Determining tree height and crown diameter from high-resolution UAV imagery. Int. J. Remote Sens. 2017, 38, 2392–2410. [Google Scholar] [CrossRef]
  7. Csillik, O.; Cherbini, J.; Johnson, R.; Lyons, A.; Kelly, M. Identification of citrus trees from unmanned aerial vehicle imagery using convolutional neural networks. Drones 2018, 2, 39. [Google Scholar] [CrossRef] [Green Version]
  8. Ampatzidis, Y.; Partel, V. UAV-Based High Throughput Phenotyping in Citrus Utilizing Multispectral Imaging and Artificial Intelligence. Remote Sens. 2019, 11, 410. [Google Scholar] [CrossRef] [Green Version]
  9. Wu, J.; Yang, G.; Yang, H.; Zhu, Y.; Li, Z.; Lei, L.; Zhao, C. Extracting apple tree crown information from remote imagery using deep learning. Comput. Electron. Agric. 2020, 174, 105504. [Google Scholar] [CrossRef]
  10. Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  11. Redmon, J.; Farhadi, A. YOLO9000: Better, Faster, Stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 6517–6525. [Google Scholar]
  12. Luo, Z.; Yu, H.; Zhang, Y. Pine Cone Detection Using Boundary Equilibrium Generative Adversarial Networks and Improved YOLOv3 Model. Sensors 2020, 20, 4430. [Google Scholar] [CrossRef] [PubMed]
  13. Gai, R.; Chen, N.; Yuan, H. A detection algorithm for cherry fruits based on the improved YOLO-v4 model. Neural Comput. Appl. 2021, 5, 1–12. [Google Scholar]
  14. Tian, Y.; Yang, G.; Wang, Z.; Wang, H.; Li, E.; Liang, Z. Apple detection during different growth stages in orchards using the improved YOLO-V3 model. Comput. Electron. Agric. 2019, 157, 417–426. [Google Scholar] [CrossRef]
  15. Zhao, H.; Shi, J.; Qi, X.; Wang, X.; Jia, J. Pyramid Scene Parsing Network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  16. Sun, Z.Y.; Chen, Y.J.; Yang, L.; Yang, L.; Tang, G.L.; Yuan, S.X.; Lin, Z.W. Low altitude remote sensing of light and small UAV and its application in ecology. J. Appl. Ecol. 2017, 28, 528–536. [Google Scholar]
  17. Han, Y.J.; Ding, N.; Feng, H.K.; Zhang, C.L. Remote Sensing Estimation of Nitrogen Content in Winter Wheat Leaves Based on Random Forest Algorithm. Mod. Agric. Sci. Technol. 2019, 13, 5. [Google Scholar]
  18. Zhou, Z.; Sa, J.M.; Fan, Y.X.; Shuai, C.; Gao, S. Remote Sensing Land Usage Classification and Landscape Pattern Analysis Based on Random Forest. Comput. Syst. Appl. 2020, 29, 44–52. [Google Scholar]
  19. Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  20. Ashraf, S.; Ahmed, T. Machine Learning Shrewd Approach for an Imbalanced Dataset Conversion Samples. J. Eng. Technol. 2020, 11, 1. [Google Scholar]
  21. Kaur, P.; Pannu, H.S.; Malhi, A.K. Plant disease recognition using fractional-order Zernike moments and SVM classifier. Neural Comput. Appl. 2019, 31, 8749–8768. [Google Scholar] [CrossRef]
  22. Mia, M.R.; Chhoton, A.C.; Mozumder, M.H.; Hossain, S.A.; Hossan, A. An Approach for Mango Disease Recognition using K-Means Clustering and SVM Classifier. In Proceedings of the 8th International Conference System Modeling and Advancement in Research Trends (SMART), Moradabad, India, 22–23 November 2019. [Google Scholar]
  23. Sannakki, S.S.; Rajpurohit, V.S.; Nargund, V.B. SVM-DSD: SVM Based Diagnostic System for the Detection of Pomegranate Leaf Diseases. In Proceedings of the Advances in Intelligent Systems and Computing; Springer Science and Business Media LLC: Berlin/Heidelberg, Germany, 2013; Volume 174, pp. 715–720. [Google Scholar] [CrossRef]
  24. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; Volume 6, pp. 770–778. [Google Scholar]
  25. Lang, Q.; Zhiyong, Z.; Longsheng, C.; Hong, S.; Minzan, L.; Li, L.; Junyong, M. Detection of Chlorophyll Content in Maize Canopy from UAV Imagery. IFAC Pap. 2019, 52, 330–335. [Google Scholar] [CrossRef]
  26. Apolo-Apolo, O.E.; Martínez-Guanter, J.; Egea, G.; Raja, P.; Pérez-Ruiz, M. Deep learning techniques for estimation of the yield and size of citrus fruits using a UAV. Eur. J. Agron. 2020, 115, 126030. [Google Scholar] [CrossRef]
  27. Modica, G.; Messina, G.; Luca, G.D.; Fiozzo, V.; Praticò, S. Monitoring the vegetation vigor in heterogeneous citrus and olive orchards. A multiscale object-based approach to extract trees’ crowns from UAV multispectral imagery. Comput. Electron. Agric. 2020, 175, 105500. [Google Scholar] [CrossRef]
  28. Osco, L.P.; Nogueira, K.; Ramos, A.P.; Pinheiro, M.M.; Furuya, D.E.; Gonçalves, W.N.; de Castro Jorge, L.A.; Junior, J.M.; dos Santos, J.A. Semantic segmentation of citrus-orchard using deep neural networks and multispectral UAV-based imagery. Precis. Agric. 2021, 4, 1–18. [Google Scholar]
  29. Fang, J.; Pan, F.J.; Lan, Y.B.; Lu, L.Q.; Cao, D.L.; Yang, D.J.; Wen, Y.T. Wheat lodging area extraction using UAV visible light remote sensing and feature fusion. Trans. Chin. Soc. Agric. Eng. 2021, 37, 73–80. [Google Scholar]
  30. Song, Z.; Zhang, Z.; Yang, S.; Ding, D.; Ning, J. Identifying sunflower lodging based on image fusion and deep semantic segmentation with UAV remote sensing imaging. Comput. Electron. Agric. 2020, 179, 105812. [Google Scholar] [CrossRef]
Figure 1. Remote sensing image of the experimental land and field enlarged picture.
Figure 1. Remote sensing image of the experimental land and field enlarged picture.
Sensors 21 04442 g001
Figure 2. Flowchart of the kiwifruit location and distribution prediction. The orchard image is used to recognize a single tree. The vine is separated from the background through image segmentation to prepare for the next step of predicting the distribution. A two-dimensional map with plant location and canopy distribution information is formed.
Figure 2. Flowchart of the kiwifruit location and distribution prediction. The orchard image is used to recognize a single tree. The vine is separated from the background through image segmentation to prepare for the next step of predicting the distribution. A two-dimensional map with plant location and canopy distribution information is formed.
Sensors 21 04442 g002
Figure 3. The original UAV image is clipped and labeled, and then input into the PSPnet network for training, and the training model is obtained. The image of the test set is clipped and input into the model to obtain the segmented image. Finally, the segmented image is stitched to obtain the segmented image of the remote sensing image.
Figure 3. The original UAV image is clipped and labeled, and then input into the PSPnet network for training, and the training model is obtained. The image of the test set is clipped and input into the model to obtain the segmented image. Finally, the segmented image is stitched to obtain the segmented image of the remote sensing image.
Sensors 21 04442 g003
Figure 4. Flowchart of resampling the normalization algorithm shows that the green and blue pixels in the segmented image are the kiwifruit vine and ground, respectively. After binarization, the image is clipped with multiple sampling sizes, and the proportion of the kiwifruit pixels in each image is calculated. Then, the color is filled according to the gradient order. Finally, the vine distribution map is generated.
Figure 4. Flowchart of resampling the normalization algorithm shows that the green and blue pixels in the segmented image are the kiwifruit vine and ground, respectively. After binarization, the image is clipped with multiple sampling sizes, and the proportion of the kiwifruit pixels in each image is calculated. Then, the color is filled according to the gradient order. Finally, the vine distribution map is generated.
Sensors 21 04442 g004
Figure 5. By fusing the resampling distribution image, the binarization distribution image, and the target recognition location image, the binarization location distribution image and the location distribution image are obtained.
Figure 5. By fusing the resampling distribution image, the binarization distribution image, and the target recognition location image, the binarization location distribution image and the location distribution image are obtained.
Sensors 21 04442 g005
Figure 6. Classification effect picture under the different classification methods. As demonstrated, (a) is the result of the PSPnet classification, (b) is the result of the SVM classification, and (c) is the result of the RF classification.
Figure 6. Classification effect picture under the different classification methods. As demonstrated, (a) is the result of the PSPnet classification, (b) is the result of the SVM classification, and (c) is the result of the RF classification.
Sensors 21 04442 g006
Figure 7. Comparison of the deep semantic segmentation and traditional machine learning segmentation. (a) The red part is the classification result of PSPnet, and the white pixels represent the classification effect of the SVM. It can be observed that the red pixels are more continuous and concentrated, and there are many surrounding white pixels. (b) The well covers and scaffolds in the UAV image. (c) The SVM mistakenly identifies irrelevant objects.
Figure 7. Comparison of the deep semantic segmentation and traditional machine learning segmentation. (a) The red part is the classification result of PSPnet, and the white pixels represent the classification effect of the SVM. It can be observed that the red pixels are more continuous and concentrated, and there are many surrounding white pixels. (b) The well covers and scaffolds in the UAV image. (c) The SVM mistakenly identifies irrelevant objects.
Sensors 21 04442 g007
Figure 8. Binary images obtained by the different thresholds under three classification methods.
Figure 8. Binary images obtained by the different thresholds under three classification methods.
Sensors 21 04442 g008
Figure 9. (a) The IOU curve of S = 35 obtained by the three classification methods changes with the threshold. (b) The IOU curve of S = 70 obtained by the three classification methods changes with the threshold.
Figure 9. (a) The IOU curve of S = 35 obtained by the three classification methods changes with the threshold. (b) The IOU curve of S = 70 obtained by the three classification methods changes with the threshold.
Sensors 21 04442 g009
Figure 10. Accuracy of the three segmentation methods under the different sampling scales and threshold levels: (a) PSPnet, (b) RF, and (c) SVM.
Figure 10. Accuracy of the three segmentation methods under the different sampling scales and threshold levels: (a) PSPnet, (b) RF, and (c) SVM.
Sensors 21 04442 g010
Figure 11. IoU curves of 40 test images with different sampling sizes when the threshold level T = 5.
Figure 11. IoU curves of 40 test images with different sampling sizes when the threshold level T = 5.
Sensors 21 04442 g011
Figure 12. Threshold level T = 5, sampling size S = 70 pixels, test set image true value pixel, and the predicted pixel number change curve.
Figure 12. Threshold level T = 5, sampling size S = 70 pixels, test set image true value pixel, and the predicted pixel number change curve.
Sensors 21 04442 g012
Table 1. Aerial photography and image parameter.
Table 1. Aerial photography and image parameter.
Route ParametersDetailed Data
flight altitude20 m
Heading/sideward overlap80%
Image size5472 × 3648
Image channelRGB
Imaging bandR (700 mm) G (546.1 mm) B (435.8 mm)
Table 2. Training environment and equipment.
Table 2. Training environment and equipment.
DeviceVersion
Programming frameworkpytorch1.7.1
Programing languagePython3.6.10
GPUNVIDIA-GeForce RTX3080s-OC-10GB
CPUInter-Xeon E5-2600-8-core processor
Table 3. PSPnet training hyper parameters and the formula.
Table 3. PSPnet training hyper parameters and the formula.
Hyper ParameterValue
Epoch100
Batch_size8, 4
Learning_rate 10 4 , 10 5
Loss_functioncross entropy loss
Table 4. Block color and category correspondence.
Table 4. Block color and category correspondence.
Resample Area ColorClasses
Dark greenCore area
GreenCoverage area
WhiteTransition area
BrownNaked land
Table 5. Segmented image evaluation.
Table 5. Segmented image evaluation.
MethodmAPSNR
RF75.26%7.2498
SVM85.8%10.5154
PSPnet71.2%12.9523
Table 6. Threshold level and the threshold correspondence.
Table 6. Threshold level and the threshold correspondence.
Threshold Level: T12345678
Threshold0.880.770.660.550.440.330.220.11
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Niu, Z.; Deng, J.; Zhang, X.; Zhang, J.; Pan, S.; Mu, H. Identifying the Branch of Kiwifruit Based on Unmanned Aerial Vehicle (UAV) Images Using Deep Learning Method. Sensors 2021, 21, 4442. https://doi.org/10.3390/s21134442

AMA Style

Niu Z, Deng J, Zhang X, Zhang J, Pan S, Mu H. Identifying the Branch of Kiwifruit Based on Unmanned Aerial Vehicle (UAV) Images Using Deep Learning Method. Sensors. 2021; 21(13):4442. https://doi.org/10.3390/s21134442

Chicago/Turabian Style

Niu, Zijie, Juntao Deng, Xu Zhang, Jun Zhang, Shijia Pan, and Haotian Mu. 2021. "Identifying the Branch of Kiwifruit Based on Unmanned Aerial Vehicle (UAV) Images Using Deep Learning Method" Sensors 21, no. 13: 4442. https://doi.org/10.3390/s21134442

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop