Next Article in Journal
Predictive Modeling of Surface Wear in Mechanical Contacts under Lubricated and Non-Lubricated Conditions
Next Article in Special Issue
Pedestrian Detection Using Multispectral Images and a Deep Neural Network
Previous Article in Journal
Managing Vibration Training Safety by Using Knee Flexion Angle and Rating Perceived Exertion
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Pedestrian Detection Algorithm Based on Score Fusion for Multi-LiDAR Systems

1
College of Intelligence Science and Technology, National University of Defense Technology, Changsha 410073, China
2
Science and Technology on Near-Surface Detection Laboratory, Wuxi 214035, China
*
Authors to whom correspondence should be addressed.
Sensors 2021, 21(4), 1159; https://doi.org/10.3390/s21041159
Submission received: 5 January 2021 / Revised: 27 January 2021 / Accepted: 2 February 2021 / Published: 7 February 2021
(This article belongs to the Special Issue Sensor Fusion for Autonomous Vehicles)

Abstract

:
Pedestrian detection plays an essential role in the navigation system of autonomous vehicles. Multisensor fusion-based approaches are usually used to improve detection performance. In this study, we aimed to develop a score fusion-based pedestrian detection algorithm by integrating the data of two light detection and ranging systems (LiDARs). We first evaluated a two-stage object-detection pipeline for each LiDAR, including object proposal and fine classification. The scores from these two different classifiers were then fused to generate the result using the Bayesian rule. To improve proposal performance, we applied two features: the central points density feature, which acts as a filter to speed up the process and reduce false alarms; and the location feature, including the density distribution and height difference distribution of the point cloud, which describes an object’s profile and location in a sliding window. Extensive experiments tested in KITTI and the self-built dataset show that our method could produce highly accurate pedestrian detection results in real-time. The proposed method not only considers the accuracy and efficiency but also the flexibility for different modalities.

1. Introduction

Pedestrian detection and tracking play an essential and significant role in diverse transportation applications [1], such as advanced driver-assistance systems (ADAS), video surveillance systems, and autonomous vehicles. Research in this area is actively ongoing. Many exciting approaches for pedestrian detection have been proposed using cameras [2] or light detection and ranging (LiDAR) [3]. Compared with cameras, LiDARs are becoming more popular due to their ability to generate highly accurate three-dimensional information. Studies on LiDAR-based pedestrian detection have been ongoing for several years; some model-based [4,5] and deep-learning-based [6,7,8,9,10] approaches have been proposed. Besides these, many multi-LiDAR systems aim to overcome the limited vertical resolution and field of view of a single LiDAR. Multi-LiDAR fusion-based approaches usually improve detection performance.
Fusing the measurements of each sensor has become a critical problem to overcome. There are primarily two groups of data fusion strategies: pre-classification and post-classification [11]. Measurement integrations occurring at raw data level or feature level are considered pre-classification fusion. The fusion of information after classification includes: score level, rank level, and decision level [12]. For multi-LiDAR systems, classic approaches working in raw data fusion are well-established. They have thus been extensively used for environment reconstruction [13], moving object [14], and negative obstacles [15] detection. Generally, to integrate different types of information sources, we need feature- or higher-level fusion strategies. Notably, these high-level algorithms have two significant advantages over the raw data-level representation: data compression and noise suppression. Premebia et al. [16] proposed a LiDAR- and vision-based pedestrian detection system with a fusion process occurring at the feature level. More recently, decision fusion methods have attracted more attention for lane [17] and object [18] detection. Typically, these decision-level fusion methods can provide object detection and classification [19] under a probabilistic framework [20].
These kinds of algorithms require the extrinsic calibration of different sensors [21]. Due to the autonomous vehicles navigating in a dynamic environment and LiDAR sensors providing a low data update rate, multi-LiDAR systems need careful spatial and temporal synchronization. Many motion compensation [22] algorithms have been developed to overcome the ego-motion error. However, even after elaborate calibration work, the relative movement between vehicles and pedestrians, the temporal synchronization error, and the phase lock offset between LiDARs can still produce some mismatch between raw LiDAR data.
This motivated our score-level fusion approach for the multi-LiDAR system. We propose a real-time and easy-to-deploy pedestrian detection solution based on two LiDARs. Compared with the rank or decision level, score-level fusion provides more insight into the post-classification process. With this in mind, we first rasterize the point cloud on theh x-y plane of the LiDAR coordinate system. Then, a 3D sliding window is adopted on the x-y plane to generate the proposals. We propose two new features to speed up the sliding process and reduce false alarms: the central points density feature and the location feature. These two types of features act as a filter and a coarse classifier, respectively, to reject false positives in the early stage, while the remains can be treated as candidate windows. A fine classifier adopting AdaBoost then performs on seven kinds of geometric features to obtain the scores. After that, we adopt the non-maximum suppression (NMS) process to reduce overlapping windows. The detection results from these two sub-LiDARs are then fused using the Bayesian rule. The comprehensive performance of the proposed method is evaluated by designing several experiments on KITTI and our self-built dataset.
Our proposal has three advantages: (1) it does not need precise time synchronization and motion compensation, (2) it can provide higher precision than a single LiDAR and raw-level fusion of two LiDARs, and (3) it is easily implemented and has a small computational burden. Though we only used two LiDARs to evaluate this algorithm’s efficiency, this framework is flexible and can be quickly extended to more LiDARs or cameras.
The rest of this paper is organized as follows: In Section 2, we introduce some related works. Section 3 provides an overview of the proposed approach. In Section 4, experimental results are presented in two parts. Finally, the conclusions are summarized in Section 5.

2. Related Work

LiDAR-based pedestrian detection has been a hot topic in recent years. It is an essential module in the environment perception system of autonomous vehicles.
Inspired by image-based object detection [4,5], some model-based approaches have been proposed. In these approaches, point clouds are firstly segmented based on an unsupervised clustering method, and hand-crafted features are then extracted from the object candidate to train classifiers such as support vector machine (SVM) [23]. The main disadvantage of this approach is the potential incorrect segmentation, including over- and under-segmentation. The quality of the segmentation is vital for the following classification of point clouds.
In recent years, due to their excellent performance, deep learning-based approaches have been widely used. These approaches can be mainly divided into two types. The first type represents the point cloud as 3D voxel grids [8] or 2D orthogonal images by projecting it to the 2D ground plane, and then convolutional neural networks (CNNs) can be applied. BirdNet [10] and TopNet-HighRes [24] both project the 3D point cloud to the bird’s eye view (BEV) and then adopt a deep neural network for detection. The method in Li et al. [7] projects the 3D point cloud to the front-view depth map and applies a 2D CNN on it. Ku et al. [25] and Chen et al. [26] proposed networks that combine features extracted from LIDAR views and RBG images for 3D object detection. The second type involves using the point cloud’s raw coordinates as the input and processing them through an integrated network architecture [27,28]. Complexer-You Only Look Once (YOLO) [29] can achieve real-time 3D object detection with state-of-the-art accuracy. However, the deep-learning-based approaches have a high computational cost and require specific computational devices such as graphical processing units (GPUs), hindering their practical application.
It is generally thought that denser LiDAR points may lead to better detection performance. Therefore, some researchers have tried to fuse data from multiple independent LiDARs. Extensive work has been completed in this area but it primarily concerned raw data fusion methods. Kevin et al. used multiple LiDARs to detect both obstacles and geometric features such as curbs, berms, and shoulders [30]. Mertz et al. [14] proposed a moving object detection system. They fused multiple sensors, including 2D and 3D LiDARs. Negative obstacles considerably influence autonomous vehicle safety because they are difficult to detect at very early stages [31]. Larson et al. presented a negative obstacle detector using a Velodyne HDL-64E [32]. Different from the traditional upright set up on the roof of the vehicle, Shang et al. [15] proposed a novel setup method of two LiDARs mounted with a tilt angle on two sides of the vehicle roof. With an overlap area, the LiDAR data density in front of the vehicle was considerably improved, which is beneficial for detecting negative obstacles.

3. The Proposed Approach

The overall workflow of this study design is summarized in the top half of Figure 1 while the corresponding detection procedures are graphically shown in the lower half. The method consists of three major modules: object proposal, fine classification, and score fusion.
The first module, which performs object candidates, contains a bounding box filter, a coarse classifier, and a one-class SVM. We apply two low-dimensional features in this module to improve the operation efficiency. Fine classification aims to classify further, score the selected sliding window, and eliminate overlapping windows. This module includes combined features computation, AdaBoost classification, and NMS. The last module fuses the information of the left and right LiDARs in the Bayesian framework. We use a parametric method to estimate the match scores’ conditional densities before using the Bayesian rule. A more detailed description of our approach is provided in the following section.

3.1. Object Proposal

Object proposal is an effective method to increase the computational efficiency of object detection [33]. The object proposal aims to generate object candidates, which are then passed to an object classifier [34]. Previously constructed methods usually perform the ground segmentation for the original point cloud and then cluster the non-ground points to generate object candidates [35]. However, this usually leads to under-segmentation, especially when groups of people walk together. We adopt an improved sliding window algorithm in this section to overcome this problem, including two stages: a bounding box filter and a coarse classifier.
We firstly discretize the input raw point cloud data into a 2.5D grid map on the x-y plane with a fixed resolution of 0.1 m × 0.1 m. Each corresponding grid stores all the height information in it. The grid map contains 50 m in front of the vehicle and 25 m on each side. We use a 0.7 m × 0.7 m sliding window to traverse the entire search area with a step size of 0.1 m.

3.1.1. Bounding Box Filter

To speed up the sliding process, we construct a bounding box filter using the following rules: (1) if the central cell is empty, the current window will be bypassed; (2) if the difference between the maximum height and the minimum height of the central cell is within a threshold, the points in the whole window are segmented as pedestrian candidates; and (3) we propose a central points density feature to further filter out the non-pedestrian candidates.
The central points density feature represents the relative concentration of the middle grids to the sliding window. The schematic of the feature is shown in Figure 2. The left column shows an extracted object candidate. The outer bounding box denotes a sliding window with a length and width of 0.7 m. The sliding window consists of 7 × 7 grids and the central window contains 3 × 3 grids.
Let F be the value of central points density feature, which is defined as:
F = n i j N , i , j { 3 , , 5 } ,
where N represents the number of points in the sliding window an n i j denotes the number of points falling in the central window. With Equation (2), we can filter out bounding boxes that do not efficiently meet these criteria.
N c e n t e r > 0 0.5 < Δ h < 2 F > 0.35

3.1.2. Coarse Classifier

To further improve the computational efficiency, we adopt a coarse classifier after the bounding box filter step. As most pedestrians are upright, it is imperative to consider an ideal bounding box that should keep the target in its center. The extracted point cloud should be complete and avoid including irrelevant surrounding points. We propose a statistical location feature including the density distribution and height difference distribution of the point cloud.
The schematic of the feature is shown in Figure 3. The same as for the center point density feature, we first project the points in the sliding window into 7 × 7 grids on the x-y plane. Then, the element in each grid is calculated using Equation (3).
Let D i j and Δ H i j be the density and height differences for the points in each grid, respectively; their definitions are:
D i j = n i j N , i , j { 1 , , 7 } Δ H i j = h i j max h i j min , i , j { 1 , , 7 } ,
where h i j max and h i j min indicate the height difference between the highest point and the lowest point in the corresponding grid, respectively.
Until now, we have been trying to finish the object proposal step. It can be considered an anomaly detection problem [36], treating the center position targets as positive samples. Therefore, we then feed the location feature vector into a one-class SVM [37] classifier to generate object candidates.

3.2. Fine Classification

We manually label positive samples to train AdaBoost classifiers for each sub-LiDAR. We choose to extract a combined feature, including seven kinds of geometric properties, as shown in Table 1. The set of feature values of each object candidate forms a vector f = ( f 1 , . . . , f 7 ) . Features f 1 , f 2 , and f 3 [38] describe the geometric properties of the object point cloud, which can be used to initially classify the point cloud. They are the number of points, the distance from the autonomous vehicle, and the point cloud’s maximum height in the z direction, respectively. Features f 4 and f 5 are the three-dimensional covariance matrix C and its eigenvalues, respectively. The covariance matrix is composed of six independent vectors, and the eigenvalues are arranged in descending order. The matrix C is defined as:
C = cov ( x , x ) cov ( x , y ) cov ( x , z ) cov ( y , x ) cov ( y , y ) cov ( y , z ) cov ( z , x ) cov ( z , y ) cov ( z , z ) ,
where:
cov ( x , y ) = n = 1 N x i x ¯ y i y ¯ n 1 .
Feature f 6 is an inertia tensor matrix [39], which is physically equivalent to the mass in Newtonian mechanics. It describes the overall distribution of the point cloud stably. The matrix I is defined as:
I = I x x I x y I x z I y x I y y I y z I z x I z y I z z ,
where:
I x x = n = 1 N y i 2 + z i 2 I x y = I y x = n = 1 N x i y i ,
where x, y, and z represent the 3D coordinates of each point, and N denotes the number of points of the point cloud.
Feature f 7 describes the rotational projection statistics [40], which are obtained by rotationally projecting the adjacent points of a feature point onto 2D planes and calculating a set of statistics of the distribution of these projected points.
In the training phase, we adopt a data augmentation strategy to enhance classification performance. We use a mask to randomly remove part of the point cloud at a ratio of 10%, 30%, or 50% to simulate occlusion. During the testing phase, each sub-LiDAR classifier outputs a score within [−100, 100] to show a pedestrian’s likelihood estimation.
We applied an NMS strategy to merge overlapping detections. The traditional NMS approach sorts the bounding boxes according to the detection scores. However, the object with the highest score is not necessarily the best, especially for point clouds. In our experiments, it caused a jitter of the object position and more false alarms in the detection of continuous frames. Therefore, we chose to use the number of points in the bounding box as the sorting criterion. The intersection over union (IoU) between the two object windows was used to judge whether they belong to the same object.

3.3. Score Fusion

Kittler et al. developed a theoretical framework for decision-making from multiple classifiers, which are representations derived from the same input source [41]. For the problem of classifying an input X into one of N possible classes { y 1 , y 2 , , y N } based on M different classifiers, based on the Bayesian decision theory [42], the input pattern should be assigned to the class y r that maximizes the posterior probability.
Assign X y r if P y r x 1 , , x R P y k x 1 , , x R
In our scenario, we can write:
Assign X P e d if P P e d S c o r e L , S c o r e R P ¬ P e d S c o r e L , S c o r e R .
In Equation (9), S c o r e L and S c o r e R indicate the scores output by the left and right classifiers, respectively; P ( P e d | S c o r e L , S c o r e R ) specifies the probability of there being a pedestrian conditioned on the scores from the left and right classifiers.
Currently, there are three broad categories to estimate these posterior probabilities: density-based score fusion, transformation based score fusion, and classifier based score fusion [12]. When the scale of training data is relatively small, the transformation-based score fusion method can be used to achieve classification using sum, max, or min classifier combination rules. Classifier based score fusion combines all scores into a feature and uses a patteren classifier to estimate P P e d S c o r e L , S c o r e R indirectly. In this study, we use KITTI and self-build data sets, with relatively large amounts of training data, so we choose density-based sore fusion to directly estimate the posterior probability.
The posterior probability can be expressed in terms of conditional joint probability densities using the Bayes rule as follows:
P P e d S c o r e L , S c o r e R = P S c o r e L , S c o r e R P e d · P ( P e d ) P S c o r e L , S c o r e R P ¬ P e d S c o r e L , S c o r e R = P S c o r e L , S c o r e R ¬ P e d · P ( ¬ P e d ) P S c o r e L , S c o r e R .
Under the assumption of the conditional independence of two classifiers, the conditional joint probability density can be expressed as the product of the marginal conditional densities, i.e.,
P S c o r e L , S c o r e R P e d = P S c o r e L P e d · P S c o r e R P e d P S c o r e L , S c o r e R ¬ P e d = P S c o r e L ¬ P e d · P S c o r e R ¬ P e d ,
where P ( S c o r e L , R | P e d ) represents the score distribution of a pedestrian. Substituting Equations (10) and (11) into Equation (9), we obtain:
Assign X P e d if P S c o r e L P e d · P S c o r e R P e d P S c o r e L ¬ P e d · P S c o r e R ¬ P e d P ¬ P e d P P e d = η .
Density estimation generally comes in two ways, by parametric or non-parametric methods [42]. If the form of the density function is assumed to be known, we can use parametric methods to estimate the parameters. k-nearest neighbour (k-NN) or some other data-driven methods do not make any assumption about the density function. In this work, we start by assuming that the likelihood probability follows a Gaussian distribution.
To verify the Gaussian distribution assumption, we chose pedestrians and trees as two kinds of objects and collected samples at different distances. Figure 4 illustrates the score distribution of these samples collected at different distance ranges. The score histogram is shown as the blue bar. The red line represents the fitted Gaussian distribution curve. The figure shows that the score is similar to the fitted Gaussian curve, thus verifying the Gaussian distribution assumption.
To analyze the negative samples’ score distribution, 4200 non-pedestrian samples, including cars, trees, shrubs, fences, etc., were collected at different distances. The score distribution of the overall negative samples is shown in Figure 5. The figure illustrates that the score histogram is also similar to the fitted Gaussian curve. P ( S c o r e L , R | P e d ) and P ( S c o r e L , R | ¬ P e d ) can be described as:
P S c o r e L , R P e d = 1 2 π σ p o s exp S c o r e L , R μ p o s 2 2 σ p o s 2 P S c o r e L , R ¬ P e d = 1 2 π σ n e g exp S c o r e L , R μ n e g 2 2 σ n e g 2 ,
where { μ p o s , σ p o s μ n e g , σ n e g } can be estimated from the training data.
Therefore, the score fusion algorithm is applied as follows:
  • Each testing sample is given a S c o r e L and S c o r e R by the classifiers of two sub-LiDARs.
  • The scores of positive and negative samples are sent to the Gaussian distribution function.
  • According to Equation (12), by comparing the posterior probability, the sample is classified as a pedestrian or not.

4. Experimental Results

To evaluate the effectiveness of the proposed algorithms, extensive experiments were conducted. First, we evaluated our pedestrian detection approach without score fusion on the KITTI 3D object detection benchmark [43], which consists of 7481 training frames and 7518 test frames from a Velodyne 64E LiDAR. After splitting the training data into a training set (3712 frames) and a validation set (3769 frames) [26], we compared our approach with state-of-the-art pedestrian detection methods. The models were all trained on the training split and evaluated on the test split and the validation split.
Then, we evaluated the whole pipeline on our self-build data sets. Our experimental platform included a laptop equipped with a quad-core 2.3 GHz Intel i5 CPU and 8 GB of RAM.

4.1. Experiment on KITTI Data Set

Our approach was evaluated on 3D detection and BEV detection on the KITTI’s official test server. Figure 6 shows some examples of the detection results. The results were calculated according to the easy, moderate, and hard difficulty levels provided by KITTI. As shown in Table 2, our proposed method significantly outperformed previous state-of-the-art methods. Among them, AVOD [25] and Complexer-YOLO [29] use both point clouds and RGB images. BirdNet [10] and TopNet-HighRes [24] are LiDAR-only methods using convolutional neural networks (CNNs).
Our approach, based on traditional models and only taking point clouds as the input, produced more competitive results than AVOD in BEV detection and outperformed the other methods by large margins on all difficulty levels in both tasks. Our approach only requires about 0.026 s runtime per frame on a quad-core CPU. This is more than twice as fast as AVOD and Complex-YOLO and four times faster than BirdNet.
In NMS, we compared the differences in the scores generated using the proposed location features and the final classifier. The influences of each module on the detection performance was analyzed by only removing the specific part and keeping all other parts unchanged. The results illustrated in Table 3 show that by adopting our proposed filter, the processing time is significantly reduced (from 96 to 39 ms and from 32 ms to 26 ms), and the detection performance is improved. This result demonstrates that the proposed filter is effective for accurately filtering out non-pedestrian proposals.
Generating scores for proposals using the location feature in NMS could further reduce computation time while maintaining similar performance compared to directly using the final classifier.

4.2. Setup of Self-Built Data Set

Two self-built data sets were prepared for the evaluation. Data set I contains five people walking in front of a parked vehicle. The measurement range is from 0 to 50 m. The total number of frames was 1676. We manually labeled the window of each pedestrian as positive samples. Details of the labeled samples are listed in Table 4. Data set II was collected in the real road environment. The training and evaluating data were extracted on different road segments. The measurement range was also up to 50 m. Details of the samples are listed in Table 5 and some examples of different test scenarios are shown in Figure 7.

4.3. Comparison of Object Proposal

As a basis for detection, we first tested the object proposal algorithms using data set I. To perform a more accurate evaluation, we used the following criteria:
  • Over-segmentation: clusters that contain fewer than 70% of the ground truth points;
  • Under-segmentation: clusters that contain more than one object or the object is lost.
The profile in Table 6 illustrates the results of the quantitative analysis. The proposed method performed better on all indicators compared to the other approaches. Clustering performance was poor due to a large number of under-segmentation errors. Compared to the typical sliding window algorithm, the proposed method’s recall increased by an average of 0.1 within 30 m. Note that as the distance increases, the segmentation accuracy decreases due to the sparsity of the point cloud, whereas our approach maintains relatively high performance. An example of the pedestrian proposal performance of the different algorithms is shown in Figure 8. The result showed that our approach can work well in complex scenarios.

4.4. Comparison of Different Classifiers

In this section, the performance of different classification algorithms is evaluated. In traditional methods [13,14,15], the raw data of two LiDARs are fused as a whole point cloud, and a fixed threshold is set for classification.
For a more detailed evaluation, all testing samples were divided into three categories: 0–15, 15–30, and 30–50 m. To eliminate the model’s impact on the results, we used three different models as the classifiers: AdaBoost, SVM, and PointNet. The same model adopted the same parameter settings. The parameters of PointNet were as follows: batch size = 32, max. epoch = 250, learning rate = 0.001, and the optimizer was Adam.
Figure 9, Figure 10 and Figure 11 show the results of the evaluation with different classifiers at different ranges, which are presented as the receiver operating characteristic (ROC) curve. The output of the raw data fusion is shown as a reference. The results of the two sub-LiDARs are also presented. Table 7 lists their area under the curve (AUC).
The performance of the two sub-LiDARs is generally lower than that of raw data fusion. However, by fusing the results of two sub-LiDARs, the proposed approach performed better than the reference, even at ranges between 30 and 50 m, where the density of the point cloud significantly decreases. It is considered that the point cloud of the sub-LiDARs is sufficiently dense, and the score fusion algorithm can overcome the detection error of a single LiDAR.
Some typical experimental results are shown in Figure 12: the four pedestrians on the right side of the first-row image are correctly detected. Three pedestrians walking close-by are also successfully detected, as shown in the second row of Figure 12.

4.5. Comparison of Processing Speed

For algorithms applied to autonomous vehicles, another important performance criterion is the processing speed. To evaluate the proposed algorithms’ computational efficiency, we tested their runtime on 300 continuous data frames. The computational device used for the proposed method contained an Intel Core-i5 CPU and 8 GB RAM. PointNet was evaluated with a GTX 1060 GPU.
The processing times of the different object proposal algorithms are shown in Table 8. The proposed improved sliding window algorithm is the fastest, and its computation time is approximately reduced by 15 ms compared to the clustering algorithm.
The processing times of different pedestrian detection algorithms are shown in Table 9. The object proposal algorithm used here is the improved sliding window algorithm proposed. The raw data fusion method based on a fixed threshold has the fastest average calculation time among the detection algorithms. The proposed algorithm requires slightly more calculation time than the raw data fusion method. PointNet is more computationally intensive and time-consuming. In general, the proposed algorithm’s average processing time is less than 30 ms, which meets the real-time requirements of autonomous vehicles.

5. Conclusions

This paper proposed a pedestrian detection algorithm based on score fusion, achieving a reasonable balance between accuracy and efficiency. The real-time performance of sensing algorithms is a critical issue for autonomous vehicles. Suppose an autonomous vehicle cruising on a street at a speed of 60 km/h; during the 0.026 s runtime of our approach, the vehicle will travel about 0.5 m. Meanwhile, considering the maximum detection distance is more than 40 m, our approach’s real-time performance meets the requirements.
The experimental results demonstrated that our approach can achieve higher accuracies than traditional raw data fusion algorithms in most cases. The proposed framework’s flexibility allows for different kinds of classification algorithms to be employed before the fusion process. However, when the pedestrian is partially occluded, the detection accuracies obviously declined. Harsh environments, especially in specific weather conditions such as rainfall, snowfall, and particles in the air, have a definite impact on algorithm performance.
In future work, we plan to improve our approach in two aspects: we intend to improve the classifier to increase detection performance, such as by adopting more discriminative features or combining them with lightweight neural networks; and we will try to use different types and quantities of classifiers from various sensors under this framework.

Author Contributions

Conceptualization, T.W.; methodology, J.H. and L.Y.; funding acquisition, K.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by by the Foundation of Science and Technology on Near-Surface Detection Laboratory under No. 6142414180207.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in Section 4.1 are openly available in KITTI dataset at http://www.cvlibs.net/datasets/kitti/raw_data.php. The data that support the findings in Section 4.2 are available from the corresponding author, L.Y., upon reasonable request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Holland, C.A.; Hill, R.; Cooke, R. Understanding the role of self-identity in habitual risky behaviours: Pedestrian road-crossing decisions across the lifespan. Health Educ. Res. 2009, 24, 674–685. [Google Scholar] [CrossRef] [Green Version]
  2. Shashua, A.; Gdalyahu, Y.; Hayun, G. Pedestrian detection for driving assistance systems: Single-frame classification and system level performance. In Proceedings of the IEEE Intelligent Vehicles Symposium, Parma, Italy, 14–17 June 2004; pp. 1–6. [Google Scholar] [CrossRef] [Green Version]
  3. Li, Z.; Wang, K.; Li, L.; Wang, F.Y. A review on vision-based pedestrian detection for intelligent vehicles. In Proceedings of the 2006 IEEE International Conference on Vehicular Electronics and Safety, Shanghai, China, 13–15 December 2006; pp. 57–62. [Google Scholar] [CrossRef]
  4. Dalal, N.; Triggs, B. Histograms of oriented gradients for human detection. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR 2005), San Diego, CA, USA, 20–25 June 2005; Volume I, pp. 886–893. [Google Scholar] [CrossRef] [Green Version]
  5. Forsyth, D. Object detection with discriminatively trained part-based models. Computer 2014, 47, 6–7. [Google Scholar] [CrossRef]
  6. Tran, D.; Bourdev, L.; Fergus, R.; Torresani, L.; Paluri, M. Learning spatiotemporal features with 3D convolutional networks. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 13–16 December 2015; pp. 4489–4497. [Google Scholar] [CrossRef] [Green Version]
  7. Li, B.; Zhang, T.; Xia, T. Vehicle detection from 3D lidar using fully convolutional network. Robot. Sci. Syst. 2016, 12. [Google Scholar] [CrossRef]
  8. Engelcke, M.; Rao, D.; Wang, D.Z.; Tong, C.H.; Posner, I. Vote3Deep: Fast object detection in 3D point clouds using efficient convolutional neural networks. In Proceedings of the IEEE International Conference on Robotics and Automation, Singapore, 29 May–3 June 2017; pp. 1355–1361. [Google Scholar] [CrossRef] [Green Version]
  9. Prokhorov, D. A Convolutional Learning System for Object Classification in 3-D Lidar Data; IEEE Press: New York, NY, USA, 2010; Volume 21, pp. 858–863. [Google Scholar] [CrossRef]
  10. Beltrán, J.; Guindel, C.; Moreno, F.M.; Cruzado, D.; García, F.; De La Escalera, A. Birdnet: A 3d object detection framework from lidar information. In Proceedings of the 2018 21st International Conference on Intelligent Transportation Systems, Maui, HI, USA, 4–7 November 2018. [Google Scholar]
  11. Jain, A.; Nandakumar, K.; Ross, A. Score normalization in multimodal biometric systems. Pattern Recognit. 2005, 38, 2270–2285. [Google Scholar] [CrossRef]
  12. Ross, A.A. Handbook of Multibiometrics; Springer: Cham, Switzerland, 2006. [Google Scholar] [CrossRef]
  13. Tu, X.; Zhang, J.; Luo, R.; Wang, K.; Zeng, Q.; Zhou, Y.; Yu, Y.; Du, S. Reconstruction of high-precision semantic map. Sensors 2020, 20, 6264. [Google Scholar] [CrossRef]
  14. Mertz, C.; Navarro-Serment, L.E.; MacLachlan, R.; Rybski, P.; Steinfeld, A.; Suppé, A.; Urmson, C.; Vandapel, N.; Hebert, M.; Thorpe, C.; et al. Moving object detection with laser scanners. J. Field Robot. 2013, 30, 17–43. [Google Scholar] [CrossRef]
  15. Shang, E.; An, X.; Wu, T.; Hu, T.; Yuan, Q.; He, H. LiDAR Based Negative Obstacle Detection for Field Autonomous Land Vehicles. J. Field Robot. 2016, 33, 591–617. [Google Scholar] [CrossRef]
  16. Premebida, C.; Ludwig, O.; Nunes, U. LIDAR and vision-based pedestrian detection system. J. Field Robot. 2009, 26, 696–711. [Google Scholar] [CrossRef]
  17. Huang, A.S.; Moore, D.; Antone, M.; Olson, E.; Teller, S. Finding multiple lanes in urban road networks with vision and lidar. Auton. Robot. 2009, 26, 103–122. [Google Scholar] [CrossRef] [Green Version]
  18. Zhao, X.; Sun, P.; Xu, Z.; Min, H.; Yu, H. Fusion of 3D LIDAR and Camera Data for Object Detection in Autonomous Vehicle Applications. IEEE Sens. J. 2020, 20, 4901–4913. [Google Scholar] [CrossRef] [Green Version]
  19. Oh, S.I.; Kang, H.B. Object detection and classification by decision-level fusion for intelligent vehicle systems. Sensors 2017, 17, 207. [Google Scholar] [CrossRef] [Green Version]
  20. He, X.; Aloi, D.N.; Li, J. Probabilistic multi-sensor fusion based indoor positioning system on a mobile device. Sensors 2015, 15, 31464–31481. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  21. Ahmad Yousef, K.M.; Mohd, B.J.; Al-Widyan, K.; Hayajneh, T. Extrinsic calibration of camera and 2D laser sensors without overlap. Sensors 2017, 17, 2346. [Google Scholar] [CrossRef] [PubMed]
  22. Liu, W. Lidar-imu time delay calibration based on iterative closest point and iterated sigma point kalman filter. Sensors 2017, 17, 539. [Google Scholar] [CrossRef] [Green Version]
  23. Song, S.; Xiao, J. Sliding shapes for 3D object detection in depth images. In Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Cham, Switzerland, 2014; Volume 8694, pp. 634–651. [Google Scholar] [CrossRef] [Green Version]
  24. Wirges, S.; Fischer, T.; Stiller, C.; Frias, J.B. Object detection and classification in occupancy grid maps using deep convolutional networks. In Proceedings of the 2018 21st International Conference on Intelligent Transportation Systems, Maui, HI, USA, 4–7 November 2018. [Google Scholar]
  25. Ku, J.; Mozifian, M.; Lee, J.; Harakeh, A.; Waslander, S.L. Joint 3D proposal generation and object detection from view aggregation. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018. [Google Scholar]
  26. Chen, X.; Ma, H.; Wan, J.; Li, B.; Xia, T. Multi-view 3D object detection network for autonomous driving. In Proceedings of the 30th IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 6526–6534. [Google Scholar] [CrossRef] [Green Version]
  27. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. PointNet: Deep learning on point sets for 3D classification and segmentation. In Proceedings of the 30th IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 77–85. [Google Scholar] [CrossRef] [Green Version]
  28. Qi, C.R.; Yi, L.; Su, H.; Guibas, L.J. PointNet++: Deep hierarchical feature learning on point sets in a metric space. arXiv 2017, arXiv:1706.02413. [Google Scholar]
  29. Simon, M.; Amende, K.; Kraus, A.; Honer, J.; Sämann, T.; Kaulbersch, H.; Milz, S.; Gross, H.M. Complexer-YOLO: Real-Time 3D object detection and tracking on semantic point clouds. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Long Beach, CA, USA, 16–20 June 2019. [Google Scholar]
  30. Peterson, K.; Ziglar, J.; Rybski, P.E. Fast feature detection and stochastic parameter estimation of road shape using multiple LIDAR. In Proceedings of the 2008 IEEE/RSJ International Conference on Intelligent Robots and Systems, Nice, France, 22–26 September 2008; pp. 612–619. [Google Scholar] [CrossRef]
  31. Rankin, A.L.; Huertas, A.; Matthies, L.H. Night-time negative obstacle detection for off-road autonomous navigation. Unmanned Syst. Technol. IX 2007, 6561, 656103. [Google Scholar] [CrossRef]
  32. Larson, J.; Trivedi, M. Lidar based off-road negative obstacle detection and analysis. In Proceedings of the IEEE Conference on Intelligent Transportation Systems, Washington, DC, USA, 5–7 October 2011; pp. 192–197. [Google Scholar] [CrossRef] [Green Version]
  33. Zitnick, C.L.; Dollár, P. Edge boxes: Locating object proposals from edges. In Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Cham, Switzerland, 2014; Volume 8693, pp. 391–405. [Google Scholar] [CrossRef]
  34. Pinheiro, P.O.; Collobert, R.; Dollar, P. Learning to segment object candidates. Adv. Neural Inf. Process. Syst. 2015, 2015, 1990–1998. [Google Scholar]
  35. Chen, T.; Dai, B.; Wang, R.; Liu, D. Gaussian-Process-Based Real-Time Ground Segmentation for Autonomous Land Vehicles. J. Intell. Robot. Syst. Theory Appl. 2014, 76, 563–582. [Google Scholar] [CrossRef]
  36. Li, K.L.; Huang, H.K.; Tian, S.F.; Xu, W. Improving one-class SVM for anomaly detection. In Proceedings of the International Conference on Machine Learning and Cybernetics, Xi’an, China, 5 November 2003; Volume 5, pp. 3077–3081. [Google Scholar] [CrossRef]
  37. Schölkopf, B.; Platt, J.C.; Shawe-Taylor, J.; Smola, A.J.; Williamson, R.C. Estimating the support of a high-dimensional distribution. Neural Comput. 2001, 13, 1443–1471. [Google Scholar] [CrossRef]
  38. Premebida, C.; Ludwig, O.; Nunes, U. Exploiting LIDAR-based features on pedestrian detection in urban scenarios. In Proceedings of the IEEE Conference on Intelligent Transportation Systems, St. Louis, MO, USA, 4–7 October 2009; pp. 18–23. [Google Scholar] [CrossRef]
  39. Navarro-Serment, L.E.; Mertz, C.; Hebert, M. Pedestrian detection and tracking using three-dimensional LADAR data. Springer Tracts Adv. Robot. 2010, 62, 103–112. [Google Scholar] [CrossRef] [Green Version]
  40. Guo, Y.; Sohel, F.; Bennamoun, M.; Lu, M.; Wan, J. Rotational projection statistics for 3D local surface description and object recognition. Int. J. Comput. Vis. 2013, 105, 63–86. [Google Scholar] [CrossRef] [Green Version]
  41. Kittler, J.; Hater, M.; Duin, R.P. Combining classifiers. Proc. Int. Conf. Pattern Recognit. 1996, 2, 897–901. [Google Scholar] [CrossRef]
  42. Duda, R.O.; Hart, P.E.; Stork, D.G. Pattern Classification, 2nd ed.; John Wiley & Sons: New York, NY, USA, 2001. [Google Scholar]
  43. Geiger, A.; Lenz, P.; Urtasun, R. Are we ready for autonomous driving? The KITTI vision benchmark suite. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 3354–3361. [Google Scholar] [CrossRef]
Figure 1. Overview of the proposed approach.
Figure 1. Overview of the proposed approach.
Sensors 21 01159 g001
Figure 2. Illustration of the proposed central points density feature.
Figure 2. Illustration of the proposed central points density feature.
Sensors 21 01159 g002
Figure 3. Definition of the proposed location feature.
Figure 3. Definition of the proposed location feature.
Sensors 21 01159 g003
Figure 4. The score distribution of typical samples collected at different distances. The score distributions of (a) pedestrians and (b) trees.
Figure 4. The score distribution of typical samples collected at different distances. The score distributions of (a) pedestrians and (b) trees.
Sensors 21 01159 g004
Figure 5. The score distribution of the negative samples collected at different distances.
Figure 5. The score distribution of the negative samples collected at different distances.
Sensors 21 01159 g005
Figure 6. Visualization of our results on the KITTI test set. The detected pedestrians are indicated with orange 3D bounding boxes in the LiDAR view. The 3D bounding boxes are projected onto the corresponding image in the upper row.
Figure 6. Visualization of our results on the KITTI test set. The detected pedestrians are indicated with orange 3D bounding boxes in the LiDAR view. The 3D bounding boxes are projected onto the corresponding image in the upper row.
Sensors 21 01159 g006
Figure 7. Examples of test scenarios in the experiment.
Figure 7. Examples of test scenarios in the experiment.
Sensors 21 01159 g007
Figure 8. An example of the better pedestrian proposal results produced by the proposed approach. (a) A scenario from the experiment; (b) the output of the clustering algorithm. The four pedestrians walking side-by-side are clustered together. (c) The output of the normal sliding window algorithm. Note that the three pedestrians on the right are mistakenly segmented. (d) The output of the proposed algorithm: the four pedestrians are correctly segmented. The color of the different clusters is randomly chosen by the algorithms.
Figure 8. An example of the better pedestrian proposal results produced by the proposed approach. (a) A scenario from the experiment; (b) the output of the clustering algorithm. The four pedestrians walking side-by-side are clustered together. (c) The output of the normal sliding window algorithm. Note that the three pedestrians on the right are mistakenly segmented. (d) The output of the proposed algorithm: the four pedestrians are correctly segmented. The color of the different clusters is randomly chosen by the algorithms.
Sensors 21 01159 g008
Figure 9. Results of evaluations at different ranges (AdaBoost).
Figure 9. Results of evaluations at different ranges (AdaBoost).
Sensors 21 01159 g009
Figure 10. Results of evaluations at different ranges (SVM).
Figure 10. Results of evaluations at different ranges (SVM).
Sensors 21 01159 g010
Figure 11. Results of evaluations at different ranges (PointNet).
Figure 11. Results of evaluations at different ranges (PointNet).
Sensors 21 01159 g011
Figure 12. Typical experimental results on real road scenarios.
Figure 12. Typical experimental results on real road scenarios.
Sensors 21 01159 g012
Table 1. Description of features.
Table 1. Description of features.
No.DescriptionDimension
f 1 Number of points1
f 2 Distance to the object1
f 3 Maximum height difference1
f 4 Three-dimensional covariance matrix6
f 5 Three-dimensional covariance matrix eigenvalue3
f 6 The normalized moment of inertia tensor6
f 7 Rotational projection statistics135
Table 2. Evaluation on KITTI test set for pedestrians.
Table 2. Evaluation on KITTI test set for pedestrians.
Method3D Detection AP (%)BEV Detection AP (%)Times (s)
EasyModerateHardEasyModerateHard
AVOD [25]36.1027.8625.7642.5833.5730.140.08
Complex-YOLO [29]17.6013.9612.7021.4218.2617.060.06
BirdNet [10]12.258.998.0620.7315.8014.590.11
TopNet-HighRes [24]10.406.926.6319.4313.5011.930.10
Ours33.7526.6423.3449.2737.9633.830.026
Table 3. Performance on KITTI validation set for pedestrians by adopting different modules.
Table 3. Performance on KITTI validation set for pedestrians by adopting different modules.
FilterNMS3D Detection AP (%)BEV Detection AP (%)Times (s)
Location FeatureClassifier ScoreEasyModerateHardEasyModerateHard
47.5443.7237.4965.0359.1551.380.026
49.7745.5339.2065.3059.0751.460.039
47.4143.6637.4764.9358.9551.300.032
49.4045.1838.7864.2558.2950.570.096
Table 4. Conditions for object proposal algorithms evaluation.
Table 4. Conditions for object proposal algorithms evaluation.
DescriptionTotal0–15 m15–30 m30–50 m
Evaluation data8380282833922160
Table 5. Conditions for pedestrian detection algorithms evaluation.
Table 5. Conditions for pedestrian detection algorithms evaluation.
DescriptionTotalPositiveNegative
Training data793121705761
Evaluation data24,736924515,491
Table 6. Comparison of the object proposal algorithms.
Table 6. Comparison of the object proposal algorithms.
Method0–15 m15–30 m30–50 m
Over (Over Is Over-Segmentation Rate)Under (Under Is under-Segmentation Rate)RecallOverUnderRecallOverUnderRecall
Clustering0.00070.49750.501800.93660.063400.96810.0319
Normal sliding window0.06010.07040.86950.01830.09490.88680.01250.14810.8394
This Paper0.00140.00320.99540.00290.01800.97910.01060.11300.8764
Table 7. Comparison of the area under the curve (AUC) of the pedestrian detection algorithms.
Table 7. Comparison of the area under the curve (AUC) of the pedestrian detection algorithms.
Method0–15 m15–30 m30–50 m
Ada.SVMPointNetAda.SVMPointNetAda.SVMPointNet
Raw data fusion0.95210.95620.94670.95690.94320.96440.89830.80770.9246
Left LiDAR0.95430.95430.93990.94840.92940.94080.90820.79240.9082
Right LiDAR0.95100.96480.95700.95260.93630.94070.86480.76820.8803
This paper0.96000.96520.96440.96590.94940.96020.92350.81850.9255
Table 8. Comparison of the processing time of the object proposal algorithms.
Table 8. Comparison of the processing time of the object proposal algorithms.
MethodAverage Computing Time (ms)
Clustering25.65
Normal sliding window10.71
This paper10.28
Table 9. Comparison of different pedestrian detection algorithms in terms of processing time.
Table 9. Comparison of different pedestrian detection algorithms in terms of processing time.
MethodAverage Computing Time (ms)
Raw data fusion23.24
PointNet56.53
This paper25.80
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wu, T.; Hu, J.; Ye, L.; Ding, K. A Pedestrian Detection Algorithm Based on Score Fusion for Multi-LiDAR Systems. Sensors 2021, 21, 1159. https://doi.org/10.3390/s21041159

AMA Style

Wu T, Hu J, Ye L, Ding K. A Pedestrian Detection Algorithm Based on Score Fusion for Multi-LiDAR Systems. Sensors. 2021; 21(4):1159. https://doi.org/10.3390/s21041159

Chicago/Turabian Style

Wu, Tao, Jun Hu, Lei Ye, and Kai Ding. 2021. "A Pedestrian Detection Algorithm Based on Score Fusion for Multi-LiDAR Systems" Sensors 21, no. 4: 1159. https://doi.org/10.3390/s21041159

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop