Next Article in Journal
Measurement Method of Temperature of the Face Gear Rim of a Spiroid Gear
Previous Article in Journal
A Combined mmWave Tracking and Classification Framework Using a Camera for Labeling and Supervised Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

LiDAR and Deep Learning-Based Standing Tree Detection for Firebreaks Applications

1
School of Technology, Beijing Forestry University, Beijing 100083, China
2
School of Education, University of Nottingham, Nottingham NG7 2RD, UK
3
Department of Civil, Construction, and Environmental Engineering, North Dakota State University, Fargo, ND 58102, USA
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(22), 8858; https://doi.org/10.3390/s22228858
Submission received: 7 October 2022 / Revised: 1 November 2022 / Accepted: 14 November 2022 / Published: 16 November 2022
(This article belongs to the Section Radar Sensors)

Abstract

:
Forest fire prevention is very important for the protection of the ecological environment, which requires effective prevention and timely suppression. The opening of the firebreaks barrier contributes significantly to forest fire prevention. The development of an artificial intelligence algorithm makes it possible for an intelligent belt opener to create the opening of the firebreak barrier. This paper introduces an innovative vision system of an intelligent belt opener to monitor the environment during the creation of the opening of the firebreak barrier. It can provide precise geometric and location information on trees through the combination of LIDAR data and deep learning methods. Four deep learning networks including PointRCNN, PointPillars, SECOND, and PV-RCNN were investigated in this paper, and we train each of the four networks using our stand tree detection dataset which is built on the KITTI point cloud dataset. Among them, the PointRCNN showed the highest detection accuracy followed by PV-RCNN and PV-RCNN. SECOND showed less detection accuracy but can detect the most targets.

1. Introduction

1.1. Background

Forest firebreaks are open areas between forests and between forests and villages, schools, factories, etc., to prevent the spread of an expanding fire. Traditional firebreaks (all-light firebreaks) are fuel-free strips, i.e., raw earth strips, constructed on the way to the spread of a potential fire to prevent or deter it. Since the 1990s, a new type of biological firebreak has emerged, a dense planting of broad and thick evergreen broadleaf trees or other fire-resistant species in open areas, blocking sunlight and inhibiting the growth of combustible materials on the surface. The latest shade-type firebreaks, which have the advantages of both whole light and biological firebreaks, were developed by USDA Forest Service technicians [1,2,3,4,5,6].
Traditional firebreak openings are generally established using conventional tools such as brush cutters and chainsaws, which not only have safety concerns but most importantly, also have low work efficiency for large amounts of explosive material clearance. In recent years, with the development of automation, practitioners have adopted intelligent equipment that can efficiently perform firebreak opening clean-up work essentially and systematically through manual remote control such as remote-controlled belt openers. However, the existing approaches of opening up firebreaks still require the manual judgment of the location and diameter information of the trees to be removed, which significantly affects the efficiency of mechanized operations. Therefore, it is important to further advance of mechanized firebreak opening operations in the field to detect standing trees in the firebreak region and obtain information on the approximate locations and sizes of trees.
To recognize the standing trees in firebreak openings, object detection [7] has been applied in the field of autonomous driving in the direction of environmental awareness. Researchers used traditional object detection algorithms in early studies to extract image features and detect pedestrians using histogram of oriented gradients (HOG) and achieved a near 100% detection success rate on the MIT pedestrian database [8]. Later, deep learning networks have also been introduced. For instance, a regional convolutional neural network (R-CNN) [9], (TA-CNN) [10]. Other scholars have applied digital image processing techniques and deep learning-based applications to fruit localization and harvesting [11]. Tang used an improved YOLOv4-tiny model to detect oil tea fruits [12].
With the increase in arithmetic power, researchers have attempted to reconstruct the standing trees scene in three dimensions (3D), through the introduction of LiDAR. Researchers have proposed more point cloud-based deep learning networks [13,14,15,16,17,18,19]. For instance, (PiontNet) network [20], and PiontNet++ [21]. Later, Zhou further processed the point cloud data using voxelization methods [22], Wang et al. added convolutional neural networks to point cloud voxelization to improve the detection accuracy [23] and Lang et al. applied PiontPillars to identify people, cars, and cyclists in point cloud data. Cheng investigates generalized LiDAR intensity normalization and its positive impact on geometric and learning-based lane marking detection [24].

1.2. Related Work

In the field of forestry, based on the database from UAV remote sensing and machine vision, machine learning networks have been investigated such as YOLOv3 [25], YOLOv4 [26], and Faster-RCNN [27] to identify spruce and count the number of trees [28]. Some scholars have studied the role of UAV LiDAR data in improving the accuracy of tree volume calculations. Based on UAV and ground-based LiDAR data [29,30]. Wang used a non-linear mixed-effects model and a hierarchical Bayesian approach to explore the feasibility of individual tree AGB estimation for long white larch (Larch Henry) in eight plots in three different regions of the Maoershan Forest Farm, Heilongjiang [31]. The tree identification has been moving towards automation, for instance, Hernandez et al. developed a hybrid pixel- and region-based algorithm to segment the images and automatically extract individual trees in the plantation and estimate their heights and crown diameters [32]. Deep learning has been applied to RGB optical images obtained by drones to classify tree species (AlexNet, VGG-16, and ResNet-50) [33]. Tang et al. used the YOLOv5 target detection network and DBSCAN point cloud clustering method to determine the location of litchi at a distance [34].
As the images taken using cameras are heavily influenced by light, they do not give accurate location and 3D information about the trees. On the other hand, LiDAR point cloud data can directly reflect objects’ locations and 3D information and are less affected by the environments. Traditional object segmentation algorithms have been investigated to obtain 3D point clouds to segment and identify trees in forestry based on LiDAR height data [35,36]. A support vector machine (SVM) classifier has been applied to extract single trees based on airborne LiDAR (light detection and ranging) and hyperspectral data [37]. Further research has been conducted using a handheld CCD camera to take sample maps surrounding the central tree for a video-based point cloud for detecting individual tree trunks in the sample map and estimating the diameter at the breast height of each trunk [38].
However, all the above-described methods may not be able to use in real-time and are highly dependent on more accurate point cloud alignment and 3D reconstruction. Traditional methods require specific knowledge and a complex tuning process as they rely on manually designed extractors for feature extraction. Thus, each method is application-specific and may not be robust enough to be used in general conditions. On the other hand, deep learning [39] is primarily data-driven for feature extraction. It can obtain deep, dataset-specific feature representations based on learning from a large number of samples, which is a more efficient and accurate representation of the dataset, with more robust abstract features extracted, better generalization, and can be end-to-end. To date, the deep learning method has not yet been used to identify trees based on point clouds and to obtain 3D information about trees. For the first time, this study uses a point cloud deep learning method combined with LiDAR to detect standing trees and localize tree point cloud data for autonomous machine operation in firebreaks opening. Multiple deep learning networks were analyzed in this paper using the KITTI dataset [40] and showed good detection results compared to detection for pedestrians and cyclists. Accurate tree objects and their location coordinates were obtained, providing tree information and reference for subsequent unmanned equipment to open up firebreaks. This study is the first to use a deep learning approach to process ground-based 3D LiDAR point cloud data for standing tree objection detection. Unlike airborne LiDAR tree identification, which uses the canopy layer as a feature, ground-based LiDAR can obtain more geometric information by directly identifying standing tree trunks in a planar view. Datasets were created based on the KITTI dataset for training and testing. The deep learning network was debugged to adapt to the detection and recognition of standing trees. Specifically, this paper has two main contributions as detailed below:
(1)
It proposes a dataset for stumpage detection produced based on the KITTI dataset. It is also demonstrated that LiDAR point cloud data are a valid data source for standing wood detection.
(2)
A deep learning approach in combination with LiDAR is developed to achieve standing timber detection, and analyze the detection accuracy and effectiveness of four-point cloud object detection networks, including the PointPillars, PointRCNN, PVRCNN, and SECOND, on the above dataset.
The remainder of the paper is organized as follows. Section 2 presents the details of the LiDAR data combined with deep learning for stumpage detection methods; Section 3 explains the experimental results, Section 3 provides a discussion, and Section 4 concludes the paper and points out some potential future work.

2. Materials and Methods

2.1. Data Acquisition and Processing

The data for the following study is from the KITTI object detection dataset, including LiDAR point cloud data and camera image data. However, we only used the point cloud data for training and the image data to validate the results in 2D. We relabeled the 2000-point cloud data from the KITTI dataset, boxed and noted the trees in the data, and aligned the LiDAR coordinate system in the label file with the camera coordinate system based on the calibration file. We created a new target species: Tree. The data set and validation set were delineated, with a total of 2000 in the training set and 2157 in the validation set.

2.2. Methods

Deep learning has given rise to many applications in scientific research, such as object detection, semantic segmentation, and tracking of moving objects. Object detection is one of the fundamental projects in the field of computer vision. In recent years, convolutional neural network models in the field of two-dimensional images have gradually matured, and deep learning algorithms in the three-dimensional field are progressively emerging. Usually, there are two types of object detection algorithms. One is the two-step algorithm represented by RCNN, the first step is selecting the candidate frame, and the SECOND is selecting the candidate frame for classification or regression. The other is a single-stage algorithm represented by Yolo and SSD [41], which is a straightforward one-step identification and detection of the class and location of the object, completed using only a convolutional neural network. Both algorithms have advantages and disadvantages; two-step algorithms are highly accurate but tend to be slower in recognition, while single-stage algorithms are faster but less accurate. In 3D object detection, researchers often split and reorganize the two algorithms, taking advantage of both to form a new network.
In this study, PiontPillars, SECOND, and PiontRCNN were used to detect standing trees in the KITTI dataset. The most suitable algorithm for standing tree detection was selected by comparing the recognition results.
A summary of the four deep-learning networks is shown in Table 1.

2.2.1. PiontPillars

PiontPillars [42] takes a different approach to voxelising point clouds, instead transforming the point cloud data into individual point pillars that can be processed as pseudo-images. After inputting the point cloud, the point cloud is divided into a grid, and the point cloud in the same grid is considered a pillar. The net will read the number of pillars, the geometric center of the pillar, and the number of point clouds, with the number of point clouds being the specified value. Suppose the points in the pillar are less than the specified value n. In that case, they are filled to n, and if more than n, random sampling is performed to make the number n. This process completes the tensorisation of the point cloud data and then uses a simplified version of the PiontNet to perform data processing and feature extraction on the tensorized point cloud data, and then performs maximum pooling to obtain a pseudo-picture, which is used as the input to the 2D CNN part to extract the picture features further, as shown in the figure. The 3D feature map is upsampled to the same size and then concatenated to fuse the features extracted from multiple convolutions. The network structure is shown in Figure 1.

2.2.2. PiontRCNN

PiontRCNN [43] is used to detect 3D objects directly from the original point cloud and is divided into two main stages. In the first stage, PiontRCNN generates 3D proposals directly from the point cloud in a bottom-up manner, learns point-by-point features, segments the original point cloud, and then generates 3D proposals from the segmented foreground points Based on this bottom-up strategy, a large number of predefined 3D boxes in the 3D space are avoided, and the search space of the generated 3D proposals is significantly limited. The role of the stage 2 network is to combine semantic features and local areas to optimize the proposals in canonical coordinates. Because bounding boxes in 2D object detection can only provide weak supervision for semantic segmentation, the two-stage 3D object detection framework, PiontRCNN, offers high robustness and accurate detection performance. The network structure is shown in Figure 2.

2.2.3. SECOND

Yan Y et al. proposed an improved sparse convolutional network and applied it to a LiDAR-based object detection task, which significantly improved the speed of training and inference [44]. In addition, they introduced a new loss function for orientation angle with better performance than other methods. Moreover, a new data augmentation method was introduced for LiDAR-only based learning problems, which improved the convergence speed and performance. The SECOND detection model consists of three components: a voxel grid feature extractor, a sparse convolutional layer (intermediate layer), and an RPN network. The voxel grid feature extractor first groups the point cloud and selects a specific range of points as input to the feature extractor. The original point cloud is then voxel-gridded and the features of each voxel are extracted using the VFE voxel feature extraction network. It would consume a considerable amount of computational resources and time. Here, SECOND uses a submanifold convolution, which limits the sparsity of the output by the sparsity of the input data, thus significantly reducing the computational effort of subsequent convolution operations. An SSD-like network is used as the RPN, consisting of three stages, i.e., (Conv × k + BN + ReLU) × 3, and the output of each stage is then deconvoluted and upsampled to form a feature map. The network structure is shown in Figure 3.

2.2.4. PV-RCNN

Shaoshuai Shi et al. proposed a novel network structure PV-RCNN combining a point_based approach and a voxel_based approach [45].
A two-stage strategy is proposed: in order to reduce the number of voxels needed to encode the scene, the original point cloud is sampled by FPS to obtain a portion of key points to represent the original 3D scene, and then a point-based set abstraction is used to aggregate the key points. Set abstraction module is used to summarize the multi-scale point cloud information by aggregating the voxel features of the points near the key points. In this way, the whole scene can be efficiently encoded into a small set of multiscale keypoint features.
Keypoint-to-grid RoI feature abstraction step: extracting summarized grid point features from keypoints. Considering that each 3D proposal corresponds to the location of its grid points, the proposed RoI-grid pooling module uses a multiple-radius keypoint SA layer for each grid point to aggregate keypoint features with multi-scale contextual information to obtain grid point features. Then, all grid point features in the proposal can be jointly used to optimize the proposal.
The VSA module is used to encode multiscale semantic features from the 3D CNN feature body to the keypoint. Here, the SA operation proposed by PointNet++ is used to aggregate the voxel-level feature volume. That is, the points around the keypoint are now regular voxels with multiscale semantic features, rather than adjacent raw points learned from PointNet to the features. Given a range threshold, the voxels whose distance from the keypoint is less than this threshold are the neighboring voxels of the keypoint; using the PointNet block to obtain the features of the keypoint based on the features of the neighboring voxels; the voxel CNN of PV-RCNN has 4 layers, so each keypoint can obtain 4 features, and different range thresholds can be set in these four layers to aggregate the features of different sensory fields. In these 4 layers, different range thresholds can be set to aggregate the local voxel features of different sensory fields, so that richer feature information can be obtained, and finally, these 4 features are concatenated to generate the multi-scale semantic features of that key point.
For each 3D RoI, RoI-grid pooling is proposed to aggregate the key point features to the RoI grid points with multiple receptive fields. Uniformly sample 6 × 6 × 6 grid points within each 3D proposal; use set abstraction operation to find the nearest neighboring key points within radius r for each grid point and use PointNet to aggregate the features of grid points from the nearest neighboring keypoint features.
The confidence prediction branch is trained in the proposal and ground-truth boxes using the previous 3D IoU. The network structure is shown in Figure 4.

2.3. Training Strategy

We modified the original network in order to apply it to stumpage detection. In order to have a better migration effect, we removed all the original labels and added a completely new class (tree) based on the KITTI dataset, and created a configuration file specifically for this category to set up the configuration of the network.
In order to adapt the size of the model and the object to be detected, we modified the AnchorSize parameter in the configuration files of PointPillars and SECOND. AnchorSize indicates the size of the prediction frame generated by the model for different objects, different detection objects need to set different AnchorSize. Referring to the AnchorSize of Pedestrian in PointPillars [0.6, 1.76, 1.73] and combining with the experiment, we modify the AnchorSize parameter in PointPillars to [0.6, 1.76, 2.73], and set the AnchorSize parameter in SECOND to [0.6, 0.8, 1.73]. In order to improve the accuracy of the prediction, we use a score threshold to keep only those prediction frames with prediction confidence greater than the score. By setting the score threshold, we can filter most of the incorrect prediction frames and keep the correct ones.
In this study, we annotated and produced the KITTI dataset. A total of 3560 tree samples from 2000-point cloud images were included to train three object detection models: PiontPillars, PiontRCNN, and SECOND. They were validated using multiple validation data to identify tree objects detected in the point cloud data and to draw prediction boxes. We use the same dataset for training, with different network structures all identifying standing tree objects but with varying accuracy. In the process of training deep learning networks, the settings of hyperparameters have a great influence on the training results and accuracy, and the settings of network parameters in this study are shown in Table 2.
PointPillars is an improved algorithm for SECOND, so both have the same epoch of 160, but the learning rate adjustment strategies are different. PointRCNN and PV-RCNN are both RCNN-based point cloud detection algorithms, with slightly different epoch settings of 70 and 80, respectively, but both adopt the same learning rate adjustment strategy, increasing the learning rate as the step increases, reaching the maximum learning rate at nearly half of the training time, and then gradually decreasing the learning rate until the learning rate increases. Then, the learning rate is gradually reduced until the learning rate tends to zero.
All four networks are implemented in PyTorch and trained using NVIDIA RTX 5000 GPUs. In contrast to using pre-trained models, the model parameters in our experiments were trained from scratch. To prevent model overfitting, batch normalization and rejection layers are used in each 3D convolutional layer. To reduce the size of high-dimensional point cloud data, averaging pools are used in each 2D convolutional layer.

3. Results and Discussion

In this study, the accuracy of several networks was obtained by training and validating the neural network. Accuracy (P), recall (R), and average precision (AP) were used in this study. AP is the area under the accuracy and recall curves. It can be used to assess the detection model’s accuracy and analyze the detection effectiveness of individual categories. The model used in this study was a 3D object detection model. Unlike the 2D IOU, which is the intersection and merging ratio of predicted and actual frames, the 3D IOU is the intersection and merging ratio of volumes, as shown in Figure 5.
For a pair of detection boxes (det box) and real boxes (gt box), their IOU value was calculated. If it exceeds the threshold of 0.5, the object is considered to be successfully detected. If the IOU is less than the threshold of 0.5, the detection is deemed unsuccessful. The accuracy rate P is defined in Equation (1) as:
P = T P T P + F P
Recall R is defined as Equation (2) as:
R = T P T P + F N
where TP is true positive, indicating that the result is correct, i.e., the detection box and label box are of the same category, and IOU is >0.5. FP is false positive for incorrectly identified objects, and FN is true negative for undetected objects. AP is calculated by assuming that there are M positive cases in N samples, which leads to M recall values (1/M, 2/M, …). The larger the AP value, the better the model effect.
The KITTI dataset was used as the training set and test set for this study. So, the accuracy results obtained were compared by using the network to recognize two other objects. We relabeled the trees in the dataset and trained them. The final evaluation yielded accuracies close to those of pedestrians and cyclists trained by the network’s original authors. KITTI classified object detection into three modes (easy, moderate, and difficult) by the criteria of degree of occlusion and truncation. As shown in Table 3, in an easy mode, the minimum bounding box height is 40 pixels, the maximum occlusion level is fully visible, and the maximum truncation is 15%. In a moderate level, the minimum bounding box height is 25 pixels, the maximum occlusion level is partially visible, and the maximum truncation is 30%. At a difficult mode, the minimum bounding box height is 25 pixels, maximum occlusion level is difficult to see, and the maximum truncation is 50%.
Through network training and testing, Table 4 shows the average accuracy of the four networks for detecting pedestrians and identifying trees. PiontPillars identified pedestrians with an accuracy of 52.08%, 43.53%, 41.49%, while the detection results of our first produced tree dataset could achieve similar results, 49.24%, 41.32%, 40.54% under the three modes, respectively. PiontRCNN’s recognition accuracy of pedestrians was 49.43%, 41.78%, and 38.63% under the three modes, while the detection for the tree dataset we made showed better results with 53.68%, 52.90%, and 51.05%, respectively. The detection results for the SECOND of pedestrian were 51.07%, 42.56%, 37.29%, but the detection results for the tree dataset were not satisfactory, with 22.80%, 21.96%, and 20.78%, respectively. The detection results for the PV-RCNN of pedestrian were 53.28%, 52.56%, 50.82%, and the detection results for the tree dataset were comparative, with 54.29%, 47.19%, and 43.49%, respectively. It can be seen that PV-RCNN has the best migration learning and the accuracy exceeds the accuracy of pedestrian objection detection.
The final loss curves and segmentation results of the four networks are shown in Figure 6.
A common metric to evaluate the speed of an inspection task is the frame per second (FPS), which is the number of images that can be processed in one second. To compare the different algorithm FPS, we used NVIDIA RTX 5000 GPU for testing in order to compare the detection times of the four networks. The detection accuracy of several networks, i.e., mAP, was also compared. The detection times of the four models are shown in Table 5.
The FPS of each network is calculated from the detection time, and a comprehensive evaluation of the network is performed by combining each performance index. The comparison results of several models are shown in Figure 7.
The horizontal axis is how often the model processes data, i.e., how much data it can process per second, and the vertical axis is the mAP of the model, which indicates the accuracy of the prediction. The circles in the figure indicate the performance of the model on the KITTI EASY level, the squares indicate the performance of the model on the KITTI MODERATE level, and the triangles indicate the performance of the model on the KITTI HARD level, and the same color indicates the same model.
From figure, we can see that PointRCNN and PV-RCNN have higher prediction accuracy but lower data processing speed, SECOND has faster data processing speed but lower prediction accuracy, and PointPillars has more moderate prediction accuracy and data processing speed.
Figure 8 shows the detection results of the PointPillars network model, showing six high-quality objection detection results. It can be seen in Figure 8b,e that the PointPillars has a good ability to predict densely aligned targets, except for the target bounding box that matches the true value, part of the detection results marked with the true value is not given but the detection results are confirmed to be correct. Figure 8a,d are the 2D camera images of the scene corresponding to Figure 8e,d, respectively. It can be seen that the detected standing wood target is also marked with a blue 2D box. The coordinates of the target are obtained by the coordinate conversion of the detected target in the 3D point cloud after the calibration file of LiDAR and camera, and the corresponding size of the 2D detection box is generated in the corresponding left out. Figure 8c,f are the bird’s eye view of Figure 8b,d, respectively, which can show the detection results in the top view angle. Figure 8b shows that some objects with similar shapes to tree trunks such as utility poles are also detected as standing trees, which is due to the point cloud features of objects such as utility poles being similar to the tree features learned by the network. The distant tree targets are not detected, which may be because the target is too far away, the point cloud is too sparse, or the feature extraction is not sufficient to be recognized as the correct category. Another potential reason may account for that the PointPillars limits the detection distance of targets to 48 m and targets beyond that range cannot be detected.
Figure 9 shows the detection effect of the SECOND network model, showing six high-quality objection detection results. In Figure 9b, it can be seen that SECOND detects the target outside the bounding box that matches the true value, and some of the detection results marked with the true value are not given, and the detection results are confirmed to be correct, while some of the marked true values in Figure 9e are not detected. Figure 9a,d is the 2D camera images of the scenes corresponding to Figure 9b,e, respectively, and it can be seen that the detected standing wood target is also marked with a blue 2D box, and like PointPillars, the coordinates of this target are also obtained by transforming the coordinates of the detected target in the 3D point cloud through the calibration file of the LiDAR and camera and generating a 2D detection box of the corresponding size on the corresponding left side. Figure 9c,f is the bird’s eye view of Figure 9b,d, respectively, showing the detection results at the top view angle. In Figure 9b, we show that some detection results are correct without the true values because the point cloud features of these misidentified objects are similar to the tree features learned by the network, while the missed detection in d is due to the fact that the true values are marked beyond the detection range limited by the network.
Figure 10 shows the detection effect of the PointRCNN network model, showing four sheets of high-quality objection detection results. In Figure 10a,b, it can be seen that PointRCNN detects all the target bounding boxes that match the true value, while in Figure 10b, some of the labeled true values are not detected, and the missed detection in Figure 10b is due to the fact that the labeled true values are beyond the detection range marked by the white boxes limited by the network. Figure 10c,d is the 2D camera images of the scenes corresponding to Figure 10a,b, respectively, and it can be seen that the detected standing wood target is also marked by the green-colored 2D boxes. As in the previous two models, the coordinates of the target are also obtained by transforming the coordinates of the detected target in the 3D point cloud through the calibration files of the LiDAR and the camera and generating a 2D detection box of the corresponding size on the corresponding left side.
Figure 11 shows the detection result graph of the PV-RCNN network model, showing six high-quality objection detection results, Figure 11a–f is different point cloud images, the detection results of this model only mark the detection results, and not marked with the true value, but it is confirmed that the detection results are correct, and there is no missed detection, and the detection effect is good.

4. Conclusions

For the first time, this paper combines deep learning with LiDAR data to detect the opposite tree. By training four network models, the most suitable model for detecting standing trees in fire escapes was determined. This method proposed yielded better results than the traditional manual observation solutions, which not only provided better technical support for the research on intelligent devices for opening fire forest roads, but also had important practical significance for opening up intelligent firebreaks.
The experimental results showed that the model could effectively identify trees, but in practice, in order to develop vision systems for intelligent devices that open up firebreaks, the geometric information of standing trees needs to be extracted, such as tree diameter. Future work is devoted to improving the data annotation and network models, so as to further improve the accuracy of detection. Future studies in this area can also explore new vertical tree information detection algorithms, so as to provide more accurate information for the opening equipment.

Author Contributions

Conceptualization, P.C., Z.L. and X.W.; methodology, Z.L.; software, Z.L. and J.Z.; validation, Z.L and P.C.; formal analysis, Z.L.; investigation, Z.L. and Y.H.; resources, X.W. and P.C.; data curation, Z.L.; writing—original draft preparation, Z.L.; writing—review and editing, Y.H. and Z.L.; visualization, Z.L.; supervision, P.C. and Z.L.; project administration, P.C. and X.W.; funding acquisition, P.C. All authors have read and agreed to the published version of the manuscript.

Funding

This study was financially supported by the National Key R&D Program of China (Grant No. 2020YFC1511601).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Green, L.R. Fuelbreaks and Other Fuel Modification for Wildland Fire Control; US Department of Agriculture, Forest Service: Washington, DC, USA, 1977.
  2. Van Wagtendonk, J.W. Use of a Deterministic Fire Growth Model to Test Fuel Treatments. In Sierra Nevada Ecosystem Project: Final Report to Congress, Volume II; University of California-Davis, Wildland Resources Center: Davis, CA, USA, 1996; Volume 43. [Google Scholar]
  3. Agee, J.K.; Bahro, B.; Finney, M.A.; Omi, P.N.; Sapsis, D.B.; Skinner, C.N.; van Wagtendonk, J.W.; Phillip Weatherspoon, C. The Use of Shaded Fuelbreaks in Landscape Fire Management. For. Ecol. Manag. 2000, 127, 55–66. [Google Scholar] [CrossRef]
  4. Rigolot, E.; Castelli, L.; Cohen, M.; Costa, M.; Duche, Y. Recommendations for fuel-break design and fuel management at the wildland urban interface: An empirical approach in south eastern France. In Proceedings of the Institute of Mediterranean Forest Ecosystems and Forest Products Warm International Workshop, Athènes, Greece, 15–16 May 2004; pp. 131–142. [Google Scholar]
  5. Dennis, F.C. Fuelbreak Guidelines for Forested Subdivisions & Communities. Ph.D. Thesis, Colorado State University, Fort Collins, CO, USA, 2005. [Google Scholar]
  6. Mooney, C. Fuelbreak Effectiveness in Canada’s Boreal Forests: A Synthesis of Current Knowledge; FPInnovations: Vancouver, BC, Canada, 2010; 53p. [Google Scholar]
  7. Zhang, H.; Xu, M.; Zhuo, L.; Havyarimana, V. A Novel Optimization Framework for Salient Object Detection. Vis. Comput. 2016, 32, 31–41. [Google Scholar] [CrossRef]
  8. Dalal, N.; Triggs, B. Histograms of Oriented Gradients for Human Detection. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; IEEE: Piscataway, NJ, USA; pp. 886–893. [Google Scholar]
  9. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation. arXiv 2013, arXiv:1311.2524. [Google Scholar]
  10. Tian, Y.; Luo, P.; Wang, X.; Tang, X. Pedestrian Detection Aided by Deep Learning Semantic Tasks. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 5079–5087. [Google Scholar]
  11. Tang, Y.; Chen, M.; Wang, C.; Luo, L.; Li, J.; Lian, G.; Zou, X. Recognition and Localization Methods for Vision-Based Fruit Picking Robots: A Review. Front. Plant Sci. 2020, 11, 510. [Google Scholar] [CrossRef] [PubMed]
  12. Tang, Y.; Zhou, H.; Wang, H.; Zhang, Y. Fruit detection and positioning technology for a Camellia oleifera C. Abel orchard based on improved YOLOv4-tiny model and binocular stereo vision. Expert Syst. Appl. 2023, 211, 118573. [Google Scholar] [CrossRef]
  13. Li, Y.; Ma, L.; Zhong, Z.; Liu, F.; Chapman, M.A.; Cao, D.; Li, J. Deep Learning for LiDAR Point Clouds in Autonomous Driving: A Review. IEEE Trans. Neural Netw. Learn. Syst. 2021, 32, 3412–3432. [Google Scholar] [CrossRef]
  14. Liu, Z.; Zhao, X.; Huang, T.; Hu, R.; Zhou, Y.; Bai, X. TANet: Robust 3D Object Detection from Point Clouds with Triple Attention. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019. [Google Scholar]
  15. Zamanakos, G.; Tsochatzidis, L.; Amanatiadis, A.; Pratikakis, I. A comprehensive survey of LIDAR-based 3D object detection methods with deep learning for autonomous driving. Comput. Graph. 2021, 99, 153–181. [Google Scholar] [CrossRef]
  16. Zheng, W.; Tang, W.; Jiang, L.; Fu, C. SE-SSD: Self-Ensembling Single-Stage Object Detector from Point Cloud. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 19–25 June 2021; pp. 14489–14498. [Google Scholar] [CrossRef]
  17. Zheng, X.; Zhu, J. Efficient LiDAR Odometry for Autonomous Driving. IEEE Robot. Autom. Lett. 2021, 6, 8458–8465. [Google Scholar] [CrossRef]
  18. Chen, Y.; Liu, S.; Shen, X.; Jia, J. Fast Point R-CNN. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9774–9783. [Google Scholar] [CrossRef]
  19. Ye, M.; Xu, S.; Cao, T. HVNet: Hybrid Voxel Network for LiDAR Based 3D Object Detection. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 1628–1637. [Google Scholar] [CrossRef]
  20. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. PointNet: Deep Learning on Point Sets for 3D Classification and Segmentation. arXiv 2016, arXiv:1612.00593. [Google Scholar]
  21. Qi, C.R.; Yi, L.; Su, H.; Guibas, L.J. PointNet++: Deep Hierarchical Feature Learning on Point Sets in a Metric Space. arXiv 2017, arXiv:1706.02413. [Google Scholar]
  22. Zhou, Y.; Tuzel, O. VoxelNet: End-to-End Learning for Point Cloud Based 3D Object Detection. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 4490–4499. [Google Scholar]
  23. Wang, C.; Cheng, M.; Sohel, F.; Bennamoun, M.; Li, J. NormalNet: A Voxel-Based CNN for 3D Object Classification and Retrieval. Neurocomputing 2019, 323, 139–147. [Google Scholar] [CrossRef]
  24. Cheng, Y.-T.; Lin, Y.-C.; Habib, A. Generalized LiDAR Intensity Normalization and Its Positive Impact on Geometric and Learning-Based Lane Marking Detection. Remote Sens. 2022, 14, 4393. [Google Scholar] [CrossRef]
  25. Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  26. Bochkovskiy, A.; Wang, C.-Y.; Liao, H.-Y.M. YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  27. Jiang, H.; Learned-Miller, E. Face Detection with the Faster R-CNN. In Proceedings of the 2017 12th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2017), Washington, DC, USA, 30 May 2017–3 June 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 650–657. [Google Scholar]
  28. Emin, M.; Anwar, E.; Liu, S.; Emin, B.; Mamut, M.; Abdukeram, A.; Liu, T. Objection detection-Based Tree Recognition in a Spruce Forest Area with a High Tree Density—Implications for Estimating Tree Numbers. Sustainability 2021, 13, 3279. [Google Scholar] [CrossRef]
  29. Liao, K.; Li, Y.; Zou, B.; Li, D.; Lu, D. Examining the Role of UAV Lidar Data in Improving Tree Volume Calculation Accuracy. Remote Sens. 2022, 14, 4410. [Google Scholar] [CrossRef]
  30. Wang, M.; Im, J.; Zhao, Y.; Zhen, Z. Multi-Platform LiDAR for Non-Destructive Individual Aboveground Biomass Estimation for Changbai Larch (Larix olgensis Henry) Using a Hierarchical Bayesian Approach. Remote Sens. 2022, 14, 4361. [Google Scholar] [CrossRef]
  31. Sparks, A.M.; Smith, A.M.S. Accuracy of a LiDAR-Based Individual Tree Detection and Attribute Measurement Algorithm Developed to Inform Forest Products Supply Chain and Resource Management. Forests 2022, 13, 3. [Google Scholar] [CrossRef]
  32. Guerra-Hernandez, J.; Gonzalez-Ferreiro, E.; Sarmento, A.; Silva, J.; Nunes, A.; Correia, A.C.; Fontes, L.; Tomé, M.; Diaz-Varela, R. Short Communication. Using High Resolution UAV Imagery to Estimate Tree Variables in Pinus Pinea Plantation in Portugal. For. Syst. 2016, 25, eSC09. [Google Scholar] [CrossRef] [Green Version]
  33. Zhang, C.; Xia, K.; Feng, H.; Yang, Y.; Du, X. Tree Species Classification Using Deep Learning and RGB Optical Images Obtained by an Unmanned Aerial Vehicle. J. For. Res. 2021, 32, 1879–1888. [Google Scholar] [CrossRef]
  34. Wang, H.; Lin, Y.; Xu, X.; Chen, Z.; Wu, Z.; Tang, Y. A Study on Long-Close Distance Coordination Control Strategy for Litchi Picking. Agronomy 2022, 12, 1520. [Google Scholar] [CrossRef]
  35. Palenichka, R.M.; Zaremba, M.B. Scale-Adaptive Segmentation and Recognition of Individual Trees Based on LiDAR Data. In Image Analysis and Recognition; Springer: Berlin/Heidelberg, Germany, 2007; pp. 1082–1092. [Google Scholar]
  36. Mohan, M.; Leite, R.V.; Broadbent, E.N.; Wan Mohd Jaafar, W.S.; Srinivasan, S.; Bajaj, S.; Dalla Corte, A.P.; do Amaral, C.H.; Gopan, G.; Saad, S.N.M.; et al. Individual tree detection using UAV-lidar and UAV-SfM data: A tutorial for beginners. Open Geosci. 2021, 13, 1028–1039. [Google Scholar] [CrossRef]
  37. La, H.P.; Eo, Y.D.; Chang, A.; Kim, C. Extraction of Individual Tree Crown Using Hyperspectral Image and LiDAR Data. KSCE J. Civ. Eng. 2015, 19, 1078–1087. [Google Scholar] [CrossRef]
  38. Zhao, Z.; Feng, Z.; Liu, J.; Li, Y. Stand Parameter Extraction Based on Video Point Cloud Data. J. For. Res. 2021, 32, 1553–1565. [Google Scholar] [CrossRef]
  39. LeCun, Y.; Bengio, Y.; Hinton, G. Deep Learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  40. Geiger, A.; Lenz, P.; Urtasun, R. Are We Ready for Autonomous Driving? The KITTI Vision Benchmark Suite. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 3354–3361. [Google Scholar]
  41. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.-Y.; Berg, A.C. SSD: Single Shot MultiBox Detector. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016. [Google Scholar] [CrossRef] [Green Version]
  42. Lang, A.H.; Vora, S.; Caesar, H.; Zhou, L.; Yang, J.; Beijbom, O. PointPillars: Fast Encoders for Object Detection from Point Clouds. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 12689–12697. [Google Scholar]
  43. Shi, S.; Wang, X.; Li, H. PointRCNN: 3D Object Proposal Generation and Detection from Point Cloud. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  44. Yan, Y.; Mao, Y.; Li, B. SECOND: Sparsely Embedded Convolutional Detection. Sensors 2018, 18, 3337. [Google Scholar] [CrossRef] [Green Version]
  45. Shi, S.; Guo, C.; Jiang, L.; Wang, Z.; Shi, J.; Wang, X.; Li, H. PV-RCNN: Point-Voxel Feature Set Abstraction for 3D Object Detection. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 10526–10535. [Google Scholar] [CrossRef]
Figure 1. PiontPillars network overview.
Figure 1. PiontPillars network overview.
Sensors 22 08858 g001
Figure 2. PiontRCNN network overview.
Figure 2. PiontRCNN network overview.
Sensors 22 08858 g002
Figure 3. SECOND network overwork.
Figure 3. SECOND network overwork.
Sensors 22 08858 g003
Figure 4. PV-RCNN network structure.
Figure 4. PV-RCNN network structure.
Sensors 22 08858 g004
Figure 5. Three-dimensional IOU schematic.
Figure 5. Three-dimensional IOU schematic.
Sensors 22 08858 g005
Figure 6. Loss curves of the four networks.
Figure 6. Loss curves of the four networks.
Sensors 22 08858 g006
Figure 7. Model detection performance (mAP) versus speed (Hz) in KITTI dataset of our newly created class.
Figure 7. Model detection performance (mAP) versus speed (Hz) in KITTI dataset of our newly created class.
Sensors 22 08858 g007
Figure 8. Test results of PointPillars. (a) Camera image; (b) 3D view of the (a); (c) a bird’s eye view of (a); (d) camera image; (e) 3d view of the (d); (f) a bird’s eye view of (d).
Figure 8. Test results of PointPillars. (a) Camera image; (b) 3D view of the (a); (c) a bird’s eye view of (a); (d) camera image; (e) 3d view of the (d); (f) a bird’s eye view of (d).
Sensors 22 08858 g008
Figure 9. Test results of SECOND. (a) Camera image; (b) 3D view of the (a); (c) a bird’s eye view of (a); (d) camera image; (e) 3D view of (d); (f) a bird’s eye view of (d).
Figure 9. Test results of SECOND. (a) Camera image; (b) 3D view of the (a); (c) a bird’s eye view of (a); (d) camera image; (e) 3D view of (d); (f) a bird’s eye view of (d).
Sensors 22 08858 g009
Figure 10. Test results of PointRCNN. (a) Three-dimensional view of the (c); (b) Three-dimensional view of the (d); (c) camera image; (d) camera image.
Figure 10. Test results of PointRCNN. (a) Three-dimensional view of the (c); (b) Three-dimensional view of the (d); (c) camera image; (d) camera image.
Sensors 22 08858 g010
Figure 11. Test results of PV-RCNN. (a) Three-dimensional view (b) Three-dimensional view (c) Three-dimensional view (d) Three-dimensional view (e) Three-dimensional view (f) Three-dimensional view.
Figure 11. Test results of PV-RCNN. (a) Three-dimensional view (b) Three-dimensional view (c) Three-dimensional view (d) Three-dimensional view (e) Three-dimensional view (f) Three-dimensional view.
Sensors 22 08858 g011aSensors 22 08858 g011b
Table 1. Network structure and function overview.
Table 1. Network structure and function overview.
NetsLayer 1Layer 2Layer 3
PointPillarsFeature net
Converting point clouds to sparse pseudo-images
Backbone
2D convolutional processing for high-level representation
Detection head
Detect and return to 3D frame
SECONDFeatures extractor
Extraction of voxel characteristics
Sparse convolution
Sparse point cloud convolution
RPN
Detect and regress target 3D frames
PointRCNNPoint wise feature vector
Extraction of voxel characteristics
Former attractions split There are two convolutional layers composed. Classify the pointsGenerate 3D frames
Detect and regress target 3D frames
PV-RCNNVoxelization
Voxelize point cloud data
Sparse convolution
3D convolution of point clouds
RPN
Detect and regress target 3D frames
Table 2. Hyperparameter setting for four networks.
Table 2. Hyperparameter setting for four networks.
NetworksPointPillarsPointRCNNSECONDPV-RCNN
Training
Parameter
Epoch1607016080
Initial learning rate0.00020.00020.00030.001
Minimum learning rate0.00022.05 × 10−80.00031.00 × 10−7
Maximum learning rate0.00020.0020.002940.01
Batch size2442
Table 3. KITTI 3D objection detection difficulty grading standard.
Table 3. KITTI 3D objection detection difficulty grading standard.
Index
Level
DifficultModerateEasy
Pixels height>25>25>40
Degree of shadingHard to detectPartially visibleAll visible
Maximum cut-off<50%<30%<15%
Table 4. Comparison of detection results of four network models.
Table 4. Comparison of detection results of four network models.
MethodModalityTreePedestrian
EasyModerateHardEasyModerateHard
PiontPillarsLiDAR49.24%41.32%40.54%52.08%43.53%41.49%
PiontRCNNLiDAR53.68%52.90%51.05%49.43%41.78%38.63%
SECONDLiDAR22.80%21.96%20.78%51.07%42.56%37.29%
PV-RCNNLiDAR53.28%52.56%50.82%54.29%47.19%43.49%
Table 5. Detection time.
Table 5. Detection time.
Model\DegreeEasyModerateHard
PointPillars0.038 s0.039 s0.042 s
SECOND0.020 s0.021 s0.021 s
PointRCNN0.095 s0.108 s0.118 s
PV-RCNN0.0790.0890.094
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Liu, Z.; Wang, X.; Zhu, J.; Cheng, P.; Huang, Y. LiDAR and Deep Learning-Based Standing Tree Detection for Firebreaks Applications. Sensors 2022, 22, 8858. https://doi.org/10.3390/s22228858

AMA Style

Liu Z, Wang X, Zhu J, Cheng P, Huang Y. LiDAR and Deep Learning-Based Standing Tree Detection for Firebreaks Applications. Sensors. 2022; 22(22):8858. https://doi.org/10.3390/s22228858

Chicago/Turabian Style

Liu, Zhiyong, Xi Wang, Jiankai Zhu, Pengle Cheng, and Ying Huang. 2022. "LiDAR and Deep Learning-Based Standing Tree Detection for Firebreaks Applications" Sensors 22, no. 22: 8858. https://doi.org/10.3390/s22228858

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop