Next Article in Journal
Change of Objectively-Measured Physical Activity during Geriatric Rehabilitation
Previous Article in Journal
Gait Phase Classification and Assist Torque Prediction for a Lower Limb Exoskeleton System Using Kernel Recursive Least-Squares Method
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Approach to Segment and Track-Based Pedestrian Detection from Four-Layer Laser Scanner Data

1
Beijing Key Lab of Urban Intelligent Traffic Control Technology, North China University of Technology, Beijing 100144, China
2
Key Laboratory for Automotive Transportation Safety Enhancement Technology of the Ministry of Communication, Chang’an University, Xi’an 710064, China
*
Author to whom correspondence should be addressed.
Sensors 2019, 19(24), 5450; https://doi.org/10.3390/s19245450
Submission received: 27 October 2019 / Revised: 5 December 2019 / Accepted: 9 December 2019 / Published: 11 December 2019
(This article belongs to the Section Remote Sensors)

Abstract

:
Pedestrian detection is a critical perception task for autonomous driving and intelligent vehicle, and it is challenging due to the potential variation of appearance and pose of human beings as well as the partial occlusion. In this paper, we present a novel pedestrian detection method via four-layer laser scanner. The proposed approach deals with the occlusion problem by fusing the segment classification results with past knowledge integration from tracking process. First, raw point cloud is segmented into the clusters of independent objects. Then, three types of features are proposed to capture the comprehensive cues, and 18 effective features are extracted with the combination of the univariate feature selection algorithm and feature correlation analysis process. Next, based on the segment classification at individual frame, the track classification is conducted further for consecutive frames using particle filter and probability data association filter. Experimental results demonstrate that both back-propagation neural network and Adaboost classifiers based on 18 selected features have their own advantages at the segment classification stage in terms of pedestrian detection performance and computation time, and the track classification procedure can improve the detection performance particularly for partially occluded pedestrians in comparison with the single segment classification procedure.

1. Introduction

Accurate and reliable obstacle classification is an important task for the environment perception module in the autonomous vehicle system, since various participants exist in the traffic environment and the motion properties of the surrounding participants directly affect the path planning of an autonomous vehicle [1]. Pedestrians are the most vulnerable traffic elements on the road, thus a great deal of attention has been paid on pedestrian detection using exteroceptive sensors. Pedestrian detection is considered as a particularly difficult problem due to the large variation of appearance and pose of human beings [2].
Most existing pedestrian detection methods rely upon several kinds of popular sensors, such as camera, radar or laser scanner [3,4]. Each sensor has its own strengths and weaknesses. The camera has been applied extensively to model the appearance characteristic of the pedestrian intuitively, but it is hard to obtain the accurate distance information and it is susceptible to illumination changes. Radar can capture the precise spatial and motion features of the obstacles, but it is not always possible to detect the static obstacles and has poor recognition capabilities. Compared with camera and radar, laser scanner enables accurate measurements and the invariance to illumination. Thus, laser scanner is used as the primary sensor in one of the most promising sensor schemes for autonomous vehicles, while camera or radar is utilized as the secondary sensor [5].
In terms of the number of scanning layers, laser scanners can be classified into three categories, namely, 2D single-layer, 2.5D multi-layer, and 3D dozens-of-layers laser scanners [6,7,8]. 2D laser scanners, e.g., the SICK LMS-111 (SICK AG, Waldkirch, Germany), provide single-layer laser beam at a fixed pitch angle, and the sparse information from single-layer point cloud is insufficient for obstacle recognition. 3D laser scanners, e.g., the Velodyne HDL-64E (Velodyne, San Francisco, CA, USA), use dozens of layers to cover 360° horizontal field of view and generate dense point cloud for omnidirectional environment modelling. In recent years, 3D laser scanners are gaining popularity in autonomous driving and intelligent vehicles, and they are usually placed on the top of a vehicle. However, the high price and external installation of 3D laser scanners limit their commercialization and popularization. Considering the practicality, 2.5D multi-layer laser scanner might be a better choice as the primary sensor than other types of laser scanners for autonomous vehicles. The existing 2.5D multi-layer laser scanner usually has four or eight layers and it is installed on the front bumper of the vehicle. Examples of 2.5D scanners include the IBEO LUX 4L and 8L (IBEO, Hamburg, Germany).
Numerous algorithms have been proposed for pedestrian detection using laser range data. Samuel et al. [7] built a pedestrian detection system based on the point cloud information from four laser planes. Particle filter was used to achieve the observation of pedestrian random movement dynamics. Carballo et al. [9] improved the pedestrian detection accuracy by introducing two novel features, namely laser intensity variation and uniformity. Gate et al. [10] used the appearance to estimate the true outlines of the tracked target. Both the geometrical and dynamical criteria of the tracked targets were utilized for pedestrian detection. Leigh et al. [11] presented a pedestrian detection and tracking system using 2D laser scanners at leg-height. Their system integrated a joint leg tracker with local occupancy grid maps to achieve robust detection. Kim et al. [12] fully exploited the feature information from 2.5D laser scanner data and developed RBFAK classifier to improve the pedestrian detection performance and reduce the computation time. Adiaviakoye et al. [13] introduced a method for detecting and tracking a crowd of pedestrians based on accumulated distribution of consecutive laser frames. This method explored the characteristics of pedestrian crowds including the velocity and trajectory. Lüy et al. [14] proposed a pedestrian detection algorithm based on a majority voting scheme using single-layer laser scanner. The scheme calculated the recognition confidence of each hypothesis over time until a high recognition confidence is achieved. Wang et al. [15] presented a framework for current frame-based pedestrian detection using 3D point clouds. A fixed-dimensional feature vector was built for each patch to solve the binary classification task. However, in their work, the precision and recall of the pedestrian detection test were unsatisfactory. Lehtomäki et al. [16] used several geometry-based point cloud features, such as local descriptor histograms, spin images, general shape and point distribution features, to improve the pedestrian detection accuracy. Xiao et al. [17] proposed a simultaneous detection and tracking method for pedestrians using 3D laser scanner. An energy function was built to incorporate the shape and motion of the point cloud, and the points belonging to pedestrians were assigned into continuous trajectories in space-time. The methods in the above literatures mainly assume that the individual people is entirely visible or the state of legs can be tracked in the classification stage. However, due to the high chance of partial occlusion, it is hard to obtain the complete contour of the individual pedestrian.
To improve the pedestrian detection performance in case of partial occlusion, some researchers attempted to use the fusion of laser and vision. García et al. [18] processed context information to enhance the pedestrian detection performance using sensor fusion of single-layer laser scanner and computer vision. Oliveira et al. [19] performed a cooperative fusion of laser and vision for pedestrian detection based on spatial relationship of parts-based classifiers via a Markov logic network. Premebida et al. [20] trained a deformable part detector using different configurations of image and laser point cloud to make a joint decision regarding whether the detected target is a pedestrian or not. It is notable that the joint calibration of a camera and laser scanner is cost-effective, and some laser points are relatively invisible to the camera.
Motivated by the analysis of the existing works in related literatures, a novel method for pedestrian detection using 2.5D laser scanner is presented in this paper. The laser scanner sensor adopted in this study is a four-layer laser scanner, i.e., IBEO LUX 4L, which is extensively used in Advanced Driver Assistance Systems (ADAS) and autonomous vehicles. The architecture of the proposed pedestrian detection method encompasses four components: segmentation, feature extraction, segment classification, and track classification, as shown in Figure 1. The proposed method differs from other pedestrian detection methods in two aspects. First, each layer of the raw data stream is employed to find the specific properties corresponding to objects of interest, and some new features are proposed. Most features are simple single-valued features, rather than high-level complex features. Second, in order to improve the detection accuracy when the pedestrians cross with each other and the partial occlusion exists, multi-pedestrian detection based on tracking is proposed to compensate the segment classification result.
The rest of this paper is organized as follows. The details of the proposed methods are presented in Section 2. Experimental results are analyzed and discussed in Section 3. Section 4 concludes the paper.

2. Proposed Methods

2.1. Segmentation

IBEO LUX 4L laser scanner is installed on the front bumper of the test vehicle, as shown in Figure 2. The raw point cloud is segmented using mean-shift clustering algorithm [21], and the clusters which are too large or too small are discarded. The point cloud of each cluster is represented by:
P c t = { p 1 t ,   p 2 t ,   p 3 t ,   ,   p N c t }
p i t = ( x i t ,   y i t ,   z i t ,   l i t )   for   i   =   1 ,   ,   N c ,   l i t { 1 ,   2 ,   3 ,   4 }
where P c t represents a set of measurements of the c-th cluster collected at time t consisting of N c points; c is the index of the cluster; N c is the number of measurement points of the c-th cluster; p i t denotes an individual point and each point is represented by lateral coordinate x i t , longitudinal coordinate y i t , vertical coordinate z i t , layer number l i t .

2.2. Feature Extraction

2.2.1. Feature Collection

To characterize the local properties of the objects of interest, each cluster is divided into 4 layers and most features are computed in each of 4 layers. This subdivision method can provide a more flexible classification representation for the occluded object [23]. Apart from the previous features in literatures, some new features are proposed to develop the pedestrian cues. All features can be categorized as three types: number-of-points-based, geometric, statistical.
The number-of-points-based features are summarized in Table 1. Feature 6 is the number of the layers which has over five measurement points. The number of points change over the layers, thus the relationship between the number of points vs. each layer is utilized as the effective feature by fitting the number of points in each layer with a linear or a second-order quadratic equation. Feature 7 is the slope of the linear fit for the number of points vs. each layer.
Features 8 and 9 are the first- and second-order coefficients of the quadratic fit for the number of points vs. each layer, respectively [10]. Feature 10 denotes the distance between the central point of the horizontal projection and the origin of the coordinate system. Feature 11 denotes the minimum distance between the horizontal projection points and the origin of the coordinate system. Features 10 and 11 tend to be stable values for the pedestrian cluster, since the region of human leg is similar, and the number of points of the pedestrian cluster decreases as the distance between the pedestrian and the laser scanner increases.
The geometric features describe the shape of the cluster intuitively and these features are summarized in Table 2. Feature 12 from [24] is the residual sum of squares of the vertical distance between the projection points and the least-squares-fitting line which is calculated with the projection points. Feature 12 describes the proximity of the distribution of the projection points to the straight line. In the formula, pxy,i denotes the projection of the i-th point in the cluster on the horizontal plane, and pl,i denotes the intersection between vertical line from the point pxy,i to the fitted straight line and the fitted line. Features 13 to 15 measure the length, width and the area of the 2D rectangular bounding box of the scan points, respectively. Features 16 to 21 represent the density of the scan points in each layer. Features 23 to 29 are used to describe the distribution of the scan points on the legs as they represent the proximity of the shape of the scan cluster to the arc or ellipse. The horizontal projection of the scan points are connected in a curve with the adjacent points along the fitted line direction, as shown in Figure 3. Features 23 to 27 reflect the properties of this projection curve. Specifically, Features 23 and 25 measure the length and bend of the horizontal projection curve, respectively. Features 26 and 27 use the inscribed angle of the projection curve to represent the curvature [25]. The inscribed angle αi at each scan point refers to the angle between two lines connecting this point with two ends of the curve, as shown in Figure 3.
The closer the mean value of inscribed angles is to 90°, the closer the curve shape is to the circle. Features 28 and 29 are defined with the least square fitting circle of the horizontal projection cluster of the scan points. Feature 28 is the sum of the residual squares of the vertical distance between the projection points and the fitting circle to imply the closeness of the projection cluster to the circle.
The statistical features also provide the cues of cluster characteristics to distinguish the pedestrian from non-pedestrian, although the physical meanings of these features are not clear. The statistical features are summarized in Table 3. Feature 30 from [26] represents the variance of the distance between the scan points in the cluster and the centroid of the cluster in 3D coordinate system. Feature 31 represents the variance of the distance between the horizontal projection points of the cluster and the mean of the horizontal projection points. Feature 32 from [24] represents the variance of the distance between the horizontal projection points of the cluster and the median of the horizontal projection points. Features 33 to 35 from [26] are the square of the second-, third-, and fourth-order centre-spaces between the horizontal projection points of the cluster and the mean of the horizontal projection points, respectively.
In order to evaluate our method for pedestrian detection, we use the datasets collected from a test vehicle moving in busy street scenes. The test vehicle is equipped with multiple sensors including IBEO LUX laser scanner and two cameras. The camera images are used to manually label the cluster of point cloud. In total, 1262 positive samples and 2463 negative samples are labelled for the segmented objects. The positive samples includes the point cloud cluster of pedestrian with different contours at various ranges, and the negative samples includes various non-pedestrian objects such as the lamp post, trees, bicycles and cars, as shown in Figure 4. For each sample, a 35-dimensional feature vector is constructed with the above 35 features, and the label is assigned with either 1 (positive) or -1 (negative). A new sample set with feature vector is finally obtained.

2.2.2. Feature Selection

After feature collection procedure, the univariate feature selection algorithm proposed by [27] is used to sort the classification failure rate for all features with a specific classifier and remove the features with high failure rate. Since feature redundancy may cause the overfitting result with low accuracy, the cross-feature correlation analysis is further combined with the univariate feature selection algorithm to remove redundant features. The details are introduced as follows.

Univariate Feature Selection Algorithm

The Back-Propagation Neural Network (BPNN) algorithm [28] is used as a specific classifier to select the effective features. Each feature is used separately to define an individual pedestrian classifier with BPNN algorithm and 35 classifiers are obtained in all. The 5-fold cross-validation is used to make sure the classifier is generalized over the labelled sample dataset, and the average classification failure rate of each feature classifier is shown in Table 4. Considering the features with high failure rate may make no sense to improve the pedestrian detection accuracy, we sorted the features by the classification failure rates of 35 classifiers and tested the classification performance of various feature sets based on the customized threshold t of the failure rate. The result is shown in Figure 5. We can see that when t = 35%, the number of the remaining features is 25 and the classification accuracy of the feature set including 25 remaining features is 0.904. When t < 35%, the number of the remaining features is less than 25 and the classification accuracy of the feature set is less than 0.904 apparently. When t > 35%, the number of the remaining features is larger than 25 while the classification accuracy of the feature set becomes gradually less than 0.904. Thus, we decided to define the features with the failure rate above 35% as invalid, and 25 remaining features are saved to analyze the feature correlation.

The Feature Correlation Analysis

Pearson correlation coefficient is used to compute the correlation between two features. It is computed by:
r ( i , j ) = cov ( f i , f j ) var ( f i ) var ( f j )
where f i and f j are two feature vectors, cov is the covariance function and var is the variance function. The closer the absolute value of r(i, j) is to 1, the higher the correlation coefficient between two features is, the larger the possibility of the redundancy between two features is. If the absolute value of r(i, j) is larger than 0.9, we consider that f i and f j are redundant. After the correlation coefficients among 25 extracted features are calculated, seven groups of feature pair are demonstrated to be redundant as shown in Table 5 and Table 6. The correlation coefficients between redundant features are represented with the italics and bold font in Table 5. If two features are redundant, the feature with the higher classification error rate is removed and the other feature with the lower error rate is kept. Thus, the redundant features pairs are arranged based on the value of correlation in descending order in Table 6, and seven features with higher classification error rate are removed. According to the classification error rate in ascending order, 18 efficient features are obtained: Feature 34, Feature 1, Feature 15, Feature 16, Feature 10, Feature 32, Feature 27, Feature 5, Feature 21, Feature 25, Feature 14, Feature 26, Feature 30, Feature 19, Feature 9, Feature 20, Feature 17 and Feature 2.

2.3. Segment Classification

In order to select the proper classifier for pedestrian segment at individual frame, we need to evaluate the effectiveness of the segment classification algorithm and answer the following questions:
(1)
What is the result of the proposed algorithm?
(2)
What is the time efficiency of the proposed algorithm?
(3)
Whether is the feature selection step effective?
(4)
Which samples are classified wrongly? Why?
In this section, BPNN classifiers are built for pedestrian detection based on the original 35 features, 25 features from univariate feature selection procedure and 18 features from the feature correlation analysis procedure, respectively. Adaboost classifier [29] is also employed based on these feature sets to compare the performance of different classifiers comprehensively. The training and validation dataset are the same as the dataset in the univariate feature selection procedure.

2.4. Track Classification

2.4.1. Tracking Process

In the dynamic process of pedestrian detection, the measurement noise of laser scanner and the point cloud variation of the pedestrian affects the performance of pedestrian detection. Pedestrian tracking can improve the accuracy of pedestrian detection with the motion information and overcome the missed classification caused by partial occlusion. Probability Data Association Filter (PDAF) [30] is a frequently-used tracking algorithm due to easy-to-use and real-time performance, but it has two disadvantages. First, the tracking gate size cannot change adaptively. Second, if two trajectories intersect or they are too close, the false association will occur and cause the trajectory of the tracked target shift or overlap. In order to overcome these disadvantages, we proposed the particle filter-based PDAF algorithm for multi-pedestrian tracking. On one hand, the nonlinear motion characteristics of particle filter satisfy the variable and random motion state of pedestrian [31]. On the other hand, the propagation range of particles in particle filter algorithm can be used to adjust the tracking gate size of PDAF adaptively to match the prediction and measurement of the track, as shown in Figure 6. The global scheme of the tracking algorithm based on particle filter and PDAF is shown in Figure 7.
The steps of the proposed tracking algorithm are as follows:
Step 1.
Initialize the particle swarm. The particle swarms for each target are the random points within a certain circular space which is centered on this target.
Step 2.
The prediction observation Z ^ ( k | k - 1 ) and covariance P ( k | k - 1 ) for each target are calculated using state equation and observation equation. The observation prediction for each particle swarm is computed at the same time.
Step 3.
The minimum bounding ellipse of the prediction particle swarm is calculated and the ellipse is used as the tracking gate threshold of PDAF to determine the valid observations Z i ( k ) , i = 1 , 2 , m ( k ) in the tracking gate. Then the associated probability between the effective observations and the target is calculated. The state estimation x ^ ( k | k - 1 ) and covariance P ( k | k ) of the target is updated based on the associated probability.
Step 4.
The state estimation x ^ ( k | k - 1 ) of PDAF at the current frame is taken as the observation value of the particle filter. The weight w ~ k i of the particle swarm and the target state estimation x ^ k are updated.
Step 5.
Particle resampling. The estimated state x ^ k in the particle filter is used as the final state at the current frame and returned to the PDAF. Perform Step 2 for the next frame.

2.4.2. Track Classification

Normally the detected object is determined directly as a pedestrian if it is classified as pedestrian at the current frame in the segment classification stage. When the tracked object is classified as non-pedestrian at the current frame in the segment classification stage, the classification result of the track at the last frame is used as a measure of whether the detected object is a human. At the same time, the velocity range of the track is also taken into consideration and it need satisfy the velocity constraint of the normal human. The mean velocity of the track at the last 3 frames can be obtained through the tracking process, and the velocity threshold for normal walking or running of most human beings is determined experimentally.

3. Experiments and Results

3.1. Segment Classification Results

To test the effectiveness of each classifier with different feature sets, Receiver Operating Characteristic (ROC) curves for each classifier are shown in Figure 8, and the results are summarized in Table 7 in terms of the accuracy, the AUC (area under the ROC) and the computation time the classifier takes for every 100 samples. The larger AUC is, the better the classification performance is. In Table 7, the accuracy of AdaBoost classifier based on 35 features reaches the highest up to 94.5%. It proves that AdaBoost algorithm does not cause overfitting with the increasing features, and 35 features we proposed are effective for pedestrian detection based on four-layer laser scanner. The accuracy of BPNN based on 35 features is the lowest, only 81.6%. After the feature selection procedure removes 17 features with high error rate and redundancy, the accuracy of BPNN based on 18 features increases to 90.7% significantly, which is close to accuracy of AdaBoost classifier. It shows that BPNN classifier is sensitive to the redundant features, and BPNN classifier based on the remaining 18 features can still obtain high accuracy and the removed features at the feature selection step contain little valid information. Overall, although the accuracy of AdaBoost classifier is higher than BPNN classifier, AdaBoost classifier takes more computation time. Note that the accuracy of AdaBoost classifier based on 35 features is only 1.3% higher than AdaBoost classifier based on 18 features, while the computation time increases by 53%. Thus, in terms of the segment classification performance and computation time, AdaBoost and BPNN classifier based on 18 features have their own advantages for pedestrian detection using multi-layer laser scanner.
The above analysis shows that the pedestrians can be classified correctly with AdaBoost and BPNN classifier based on 18 features in most cases. In the real road experiment, both pedestrians and non-pedestrians in the scene are correctly distinguished by BPNN classifier, as shown in Figure 9. In this figure, pedestrians are denoted by red circles, and non-pedestrians are depicted by black rectangles. After comparing the scene image with the detection results of the point cloud cluster, we find that most failure cases are caused by long range and the occlusion. As shown in Figure 10, the pedestrian surrounded by the blue curves in the image is wrongly identified in the point cloud view, since this pedestrian is occluded by the one surrounded by red curves. The occlusion, which often happens in real traffic scene, decreases the performance of the pedestrian classifier and eventually leads to misclassification. Pedestrians walking with sundries may also cause false detection. In Figure 10, the pedestrian surrounded by the blue curves in the image carried a carton, which is flush with his legs and interferes the laser scan on his legs, so that the point cloud returned from this pedestrian’s legs is abnormal in the point cloud top-view, and this pedestrian is wrongly classified as non-pedestrian.

3.2. Track Classification Results

To demonstrate the performance of the proposed track classification algorithm under the influence of the occlusion and the cross of pedestrian motion trajectories, real road experiments were further carried out. The parameters of the proposed algorithm were debugged to the optimal value through a large number of tests. The initial range of the particle swarm is set as a circle with a radius of 1.39 meters centred on the pedestrian’s initial position, and the number of particles in the particle swarm is 45. The trajectory similarity threshold in trajectory-cross management is 0.35 and the number of multi-target tracker is 10.
As shown in Figure 11, the pedestrians in the test scenario are numbered 1 to 5, Pedestrians 1–3 gradually walked away while Pedestrians 4 and 5 were coming. It is clear that Pedestrian 1 and 5 were very close during the walking process and their trajectories crossed. The temporary occlusion leads to the pedestrian recognition failure as described in Section 2.4.1.
The tracking trajectories of five pedestrians are shown in Figure 12. Figure 12a shows the distribution of the point cloud at the last frame and each pedestrian’s “dragging tail” is the tracking trajectory for 100 frames. It can be seen that all the tracking trajectories are continuous and smooth, and the proposed tracking algorithm can still accurately track multiple pedestrians, even if the temporary occlusion occurs for several frames.
Figure 12b,c enlarge the local areas of the tracking trajectory in Figure 12a to show the details, and the observations of the pedestrian positions from the pedestrian detection algorithm are denoted by the pink dots (Take the mean values of the horizontal and vertical coordinates of the pedestrian point cloud cluster as the position observation). Figure 12b shows a part of tracking trajectories of Pedestrians 4 and 5 at the tracking start time. It can be seen that the pedestrian recognition algorithm works well and the traces are continuous. As Pedestrians 4 and 5 become occluded by Pedestrian 2, the shape of the point cloud cluster dramatically changes, and the observations of the pedestrian position fluctuate greatly. However, the position fluctuations of Pedestrians 4 and 5 in the trajectory curve are significantly removed by the tracking algorithm. It means that the proposed tracking algorithm, which uses the propagation range of particle filter to adaptively determine the tracking gate for PDA, can enhance the accuracy and reliability of pedestrian position estimations.
Figure 12c shows a part of the trajectories when Pedestrians 1 and 5 intersects. We can see that Pedestrian 1 is not classified by the pedestrian recognition algorithm during some frames and the traces of the recognition results are interrupted due to the occlusion of Pedestrian 5 (corresponding to Scene 3 in Figure 11). However, the pedestrian tracking trajectory has never stopped. Thus the trajectory-cross management procedure is effective and it can overcome the false detection problem caused by the occlusion. When Pedestrians 1 and 5 are so close that their trajectories cross, the data association result and the tracking trajectories are still correct. Therefore, the trajectory-cross management method based on the trajectory similarity is effective to solve the trajectory aggregation and migration in the PDAF tracking method.
To evaluate the performance of the track classification method quantitatively, we collected a dataset including 1000 pedestrian tracks and 2000 non-pedestrian tracks. The tracks are the temporal series of the labelled point cloud segments. Each track has at least eight segments to ensure that the tracked object is observed for at least one second. 250 inconsistent tracks caused by the occlusions exist in the dataset. The dataset is divided into training and test sets at random. The pedestrian detection result is shown in Table 8. We can see that the track classification method based on BPNN or Adaboost classifier at successive frames achieves better performance than the individual segment classification method. Thus, the tracking process facilitates the pedestrian detection performance.
Due to the sparsity of the point cloud at long range, the pedestrian range affects the pedestrian detection performance. Figure 13 shows the quantitative performance analysis of the proposed track classification algorithm based on Adaboost classifier and 18 features at various ranges from the laser scanner sensor to the pedestrian. We can see that the best performance is between 5 m and 25 m, and good performance up to 35 m, beyond which the pedestrian detection performance declines significantly.

4. Conclusions

This study developed a novel pedestrian detection algorithm via a four-layer laser scanner. First, the raw point cloud is clustered into disjunctive segments, and three types of the features are developed including number-of-points-based, geometric and statistical features. Then, the univariate feature selection and feature correlation analysis procedures are conducted to select the effective features and remove the redundant features. Based on the segment classification, the particle filter and PDAF are combined to achieve the track classification to reduce false classification caused by the occlusion. In the road experiment, the laser scanner was mounted on an actual vehicle to collect the point cloud of surrounding environment, and the performance of the proposed pedestrian detection method is tested. Experimental results demonstrated that AdaBoost and BPNN classifiers based on 18 features have their own advantages for pedestrian detection using multi-layer laser scanner in terms of the detection performance and computation time. Moreover, the proposed pedestrian detection method based on segment and track classification using 18 features is effective even when the temporary occlusion among the pedestrians occurs.

Author Contributions

M.Z. designed this study and wrote the manuscript. R.F. designed the manuscript. W.C. and L.W. conducted the experiments. Y.M. revised this manuscript. M.Z. and W.C. discussed the experimental results together.

Funding

This work is supported by the National Key R&D Program of China under Grant No. 2018YFB1600500, National Natural Science Foundation of China under Grant Nos. 51905007, 51775053, 61603004, the Great Wall Scholar Program under Grant CIT&TCD20190304, Ability Construction of Science, Beijing Key Lab Construction Fund under Grant PXM2017-014212-000033 and NCUT start-up fund.

Conflicts of Interest

The authors declare that they have no competing interest.

References

  1. Gong, Z.; Lin, H.; Zhang, D.; Luo, Z.; Zelek, J.; Chen, Y.; Wang, C.; Li, J. A Frustum-Based Probabilistic Framework for 3D Object Detection by Fusion of LiDAR and Camera. ISPRS J. Photogramm. Remote Sens. 2020, 159, 90–100. [Google Scholar] [CrossRef]
  2. Yan, P.; Zhuo, L.; Li, J.; Zhang, H.; Zhang, J. Pedestrian Attributes Recognition in Surveillance Scenarios Using Multi-Task Lightweight Convolutional Neural Network. Appl. Sci. 2019, 9, 4182. [Google Scholar] [CrossRef] [Green Version]
  3. Fu, X.; Yu, R.; Zhang, W.; Wu, J.; Shao, S. Delving deep into multiscale pedestrian detection via single scale feature maps. Sensors 2018, 18, 1063. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Zhang, S.; Wang, C.; He, Z.; Li, Q.; Lin, X.; Zhang, J.; Yang, C.; Li, J. Vehicle Global 6-DoF Pose Estimation Under Traffic Surveillance Camera. ISPRS J. Photogramm. Remote Sens. 2020, 159, 114–128. [Google Scholar] [CrossRef]
  5. Zhang, M.; Fu, R.; Morris, D.; Wang, C. A Framework for Turning Behavior Classification at Intersections Using 3D LIDAR. IEEE Trans. Veh. Technol. 2019, 68, 7431–7442. [Google Scholar] [CrossRef]
  6. Liu, K.; Wang, W.; Wang, J. Pedestrian Detection with Lidar Point Clouds Based on Single Template Matching. Electronics 2019, 8, 780. [Google Scholar] [CrossRef] [Green Version]
  7. Gidel, S.; Checchin, P.; Blanc, C.; Chateau, T.; Trassoudaine, L. Pedestrian Detection and Tracking in an Urban Environment Using a Multilayer Laser Scanner. IEEE Trans. Intell. Transp. Syst. 2010, 11, 579–588. [Google Scholar] [CrossRef]
  8. Xu, J.; Ma, Y.; He, S.; Zhu, J. 3D-GIoU: 3D Generalized Intersection over Union for Object Detection in Point Cloud. Sensors 2019, 19, 4093. [Google Scholar] [CrossRef] [Green Version]
  9. Carballo, A.; Ohya, A.; Yuta, S. People detection using range and intensity data from multi-layered Laser Range Finders. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots & Systems, Hamburg, Germany, 26–28 September 2015; pp. 5849–5854. [Google Scholar]
  10. Gate, G.; Nashashibi, F. Using targets appearance to improve pedestrian classification with a laser scanner. In Proceedings of the IEEE Intelligent Vehicles Symposium, Eindhoven, The Netherlands, 4–6 June 2008; pp. 571–576. [Google Scholar]
  11. Leigh, A.; Pineau, J.; Olmedo, N.; Zhang, H. Person tracking and following with 2d laser scanners. In Proceedings of the IEEE International Conference on Robotics and Automation, Seattle, WA, USA, 26–30 May 2015; pp. 726–733. [Google Scholar]
  12. Kim, B.; Choi, B.; Park, S.; Kim, H.; Kim, E. Pedestrian/Vehicle Detection Using a 2.5-D Multi-Layer Laser Scanner. IEEE Sens. J. 2015, 16, 400–408. [Google Scholar] [CrossRef]
  13. Adiaviakoye, L.; Patrick, P.; Marc, B.; Auberlet, J. Tracking of multiple people in crowds using laser range scanners. In Proceedings of the IEEE Conference on Intelligent Sensors, Sensor Networks and Information Processing, Singapore, 21–24 April 2014; pp. 1–6. [Google Scholar]
  14. Lüy, M.; Çam, E.; Ulamış, F.; Uzun, İ.; Akın, S. Initial Results of Testing a Multilayer Laser Scanner in a Collision Avoidance System for Light Rail Vehicles. Appl. Sci. 2018, 8, 475. [Google Scholar] [CrossRef] [Green Version]
  15. Wang, Z.; Posner, I.; Newman, P. What could move? Finding cars, pedestrians and bicyclists in 3d laser data. In Proceedings of the IEEE Conference on Robotics and Automation, Guangzhou, China, 11–14 December 2012; pp. 4038–4044. [Google Scholar]
  16. Matti, L.; Jaakkola, A.; Juha, H. Object Classification and Recognition from Mobile Laser Scanning Point Clouds in a Road Environment. IEEE Trans. Geosci. Remote Sens. 2016, 54, 1226–1239. [Google Scholar]
  17. Xiao, W.; Vallet, B.; Schindler, K.; Paparoditis, N. Simultaneous detection and tracking of pedestrian from panoramic laser scanning data. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2016, 3, 295–302. [Google Scholar] [CrossRef]
  18. García, F.; García, J.; Ponz, A.; Escalera, A.; Armingol, M. Context aided pedestrian detection for danger estimation based on laser scanner and computer vision. Expert Syst. Appl. 2014, 41, 6646–6661. [Google Scholar] [CrossRef] [Green Version]
  19. Oliveira, L.; Nunes, U.; Peixoto, P.; Silva, M.; Moita, F. Semantic fusion of laser and vision in pedestrian detection. Pattern Recognit. 2010, 43, 3648–3659. [Google Scholar] [CrossRef]
  20. Premebida, C.; Carreira, J.; Batista, J.; Nunes, U. Pedestrian detection combining RGB and dense lidar data. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Chicago, IL, USA, 14–18 September 2014; pp. 4112–4117. [Google Scholar]
  21. Hu, X.; Li, Y.; Shan, J.; Zhang, J.; Zhang, Y. Road Centerline Extraction in Complex Urban Scenes from LiDAR Data Based on Multiple Features. IEEE Trans. Geosci. Remote Sens. 2014, 52, 7448–7456. [Google Scholar]
  22. Ibeo Automotive Systems GmbH. Operating Manual Ibeo LUX 2010 Laserscanner; Ibeo Automotive Systems GmbH: Hamburg, Germany, 2010. [Google Scholar]
  23. An, J.; Choi, B.; Sim, K.B.; Kim, E. Novel intersection type recognition for autonomous vehicles using a multi-layer laser scanner. Sensors 2016, 16, 1123. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Arras, K.; Mozos, O.; Burgard, W. Using boosted features for the detection of people in 2D range data. In Proceedings of the IEEE Conference on Robotics and Automation, Roma, Italy, 10–14 April 2007; pp. 3402–3407. [Google Scholar]
  25. Xavier, J.; Pacheco, M.; Castro, D.; Ruano, A.; Nunes, U. Fast line, arc/circle and leg detection from laser scan data in a player driver. In Proceedings of the IEEE Conference on Robotics and Automation, Barcelona, Spain, 18–22 April 2005; pp. 3930–3935. [Google Scholar]
  26. Ji, Z.; Luciw, M.; Weng, J.; Zeng, S. Incremental Online Object Learning in a Vehicular Radar-Vision Fusion Framework. IEEE Trans. Intell. Transp. Syst. 2011, 12, 402–411. [Google Scholar] [CrossRef]
  27. Hua, J.; Tembe, D.; Dougherty, R. Performance of feature-selection methods in the classification of high-dimension data. Pattern Recognit. 2009, 42, 409–424. [Google Scholar] [CrossRef]
  28. Goh, A. Back-propagation neural networks for modeling complex systems. Artif. Intell. Eng. 1995, 9, 143–151. [Google Scholar] [CrossRef]
  29. Freund, Y.; Schapire, R. A decision-theoretic generalization of on-line learning and an application to boosting. J. Comput. Syst. Sci. 1997, 55, 119–139. [Google Scholar] [CrossRef] [Green Version]
  30. Bar-Shalom, Y.; Daum, F.; Huang, J. The probabilistic data association filter. IEEE Control Syst. Mag. 2009, 29, 82–100. [Google Scholar]
  31. Arulampalam, S.; Maskell, S.; Gordon, N.; Clapp, T. A tutorial on particle filters for online nonlinear/non-Gaussian Bayesian tracking. IEEE Trans. Signal Process. 2002, 50, 174–188. [Google Scholar] [CrossRef] [Green Version]
Figure 1. The architecture of the pedestrian detection algorithm.
Figure 1. The architecture of the pedestrian detection algorithm.
Sensors 19 05450 g001
Figure 2. IBEO LUX 4L laser scanner. (a) scan layers and vertical beam divergence [22]. (b) the laser scanner installed on the test vehicle.
Figure 2. IBEO LUX 4L laser scanner. (a) scan layers and vertical beam divergence [22]. (b) the laser scanner installed on the test vehicle.
Sensors 19 05450 g002
Figure 3. The diagram of the horizontal projection curve. The dashed line denotes the fitted line of the horizontal projection points.
Figure 3. The diagram of the horizontal projection curve. The dashed line denotes the fitted line of the horizontal projection points.
Sensors 19 05450 g003
Figure 4. The labelled samples for pedestrian detection test.
Figure 4. The labelled samples for pedestrian detection test.
Sensors 19 05450 g004
Figure 5. The classification performance of various feature sets based on the customized threshold of the failure rate.
Figure 5. The classification performance of various feature sets based on the customized threshold of the failure rate.
Sensors 19 05450 g005
Figure 6. The tracking gate size of PDAF algorithm varies adaptively using particle filter.
Figure 6. The tracking gate size of PDAF algorithm varies adaptively using particle filter.
Sensors 19 05450 g006
Figure 7. The tracking algorithm scheme based on particle filter and PDAF.
Figure 7. The tracking algorithm scheme based on particle filter and PDAF.
Sensors 19 05450 g007
Figure 8. ROC curves for pedestrian detection.
Figure 8. ROC curves for pedestrian detection.
Sensors 19 05450 g008
Figure 9. Pedestrian detection result in Scene 1.
Figure 9. Pedestrian detection result in Scene 1.
Sensors 19 05450 g009aSensors 19 05450 g009b
Figure 10. Pedestrian detection result in Scene 2.
Figure 10. Pedestrian detection result in Scene 2.
Sensors 19 05450 g010
Figure 11. The tracking test scenario for multiple pedestrians.
Figure 11. The tracking test scenario for multiple pedestrians.
Sensors 19 05450 g011
Figure 12. The results of tracking trajectories. (a) the tracking trajectories of multiple pedestrians in the point cloud scene. (b) the tracking trajectories in the enlarged local area A. (c) the tracking trajectories in the enlarged local area B.
Figure 12. The results of tracking trajectories. (a) the tracking trajectories of multiple pedestrians in the point cloud scene. (b) the tracking trajectories in the enlarged local area A. (c) the tracking trajectories in the enlarged local area B.
Sensors 19 05450 g012
Figure 13. The performance of the proposed track classification algorithm based on Adaboost classifier and 18 features at various ranges from the laser scanner sensor to the pedestrian.
Figure 13. The performance of the proposed track classification algorithm based on Adaboost classifier and 18 features at various ranges from the laser scanner sensor to the pedestrian.
Sensors 19 05450 g013
Table 1. Number-of-points-based features.
Table 1. Number-of-points-based features.
#ExpressionDescription
1NcNumber of points
2Nc1Number of points from the first layer
3Nc2Number of points from the second layer
4Nc3Number of points from the third layer
5Nc4Number of points from the fourth layer
6nlThe number of the layers with more than two points
7nkSlope of the linear fit of the number of scan points vs. layers
8naFirst order coefficient of the quadratic fit of the number of points vs. each layer
9nbSecond order coefficient of the quadratic fit of the number of points vs. each layer
10DThe distance between the central point of the horizontal projection and the origin of the coordinate system
11dThe minimum distance between the horizontal projection points and the origin of the coordinate system
Table 2. Geometric features.
Table 2. Geometric features.
#FormulaDescription
12 i = 1 N c ( p x y , i p ^ l , i ) 2 / N c Linearity: p ^ l , i corresponds to the fitted line
13lLength along the fitted line at the horizontal plane
14wLength along the direction perpendicular to the fitted line at the horizontal plane
15 l × w The area of the fitting rectangle at the horizontal plane
16Nc1/A1The density of the points from the first layer
17Nc2/A2The density of the points from the second layer
18Nc3/A3The density of the points from the third layer
19Nc4/A4The density of the points from the fourth layer
20A1 + A2 + A3 + A4Sum of the areas of four layers
21(A1 + A2 + A3 + A4)/4The average value of the total areas of four layers
22 Δ X 2 + Δ Y 2 Cartesian dimension
23 i = 2 N c ( p x y , i p x y , i 1 ) Length of the connected curves
24 i = 2 N c ( | p x y , i p x y , i 1 | d ¯ ) 2 / ( N c 1 ) The variance of the connected curves
25 i = 2 N c ( p x y , i p x y , i 1 ) / l Bending of the connected curves
26 i = 1 N c 2 α i / ( N c 2 ) The inscribed angle mean (IAM)
27 i = 1 N c ( α IAM ) 2 / ( N c 2 ) The inscribed angle variance
28 i = 1 N c ( P x y , i P ^ a , i ) 2 / N c Circularity: P ^ a , i corresponds to the fitted circle
29rThe radius of the fitted circle
Table 3. Statistical features.
Table 3. Statistical features.
#FormulaDescription
30 1 N c i = 1 N c | | p 3 D , i p ¯ 3 D | | Standard deviation for 3D, p ¯ 3 D is 3D centroid
31 1 N c i = 1 N c ( p x y , i p ¯ 2 D ) 2 Standard deviation for 2D, p ¯ 2 D is the mean of the horizontal projection points
32 i = 1 N c ( p x y , i D m ) 2 / N c Variance for 2D, Dm is the median of the horizontal projection points
33 i = 1 N c ( p x y , i p ¯ 2 D ) 2 / N c Second central moment
34 i = 1 N c ( p x y , i p ¯ 2 D ) 3 / N c Third central moment
35 i = 1 N c ( p x y , i p ¯ 2 D ) 4 / N c Fourth central moment
Table 4. The average classification failure rate of each feature classifier.
Table 4. The average classification failure rate of each feature classifier.
FeatureFailure RateFeatureFailure RateFeatureFailure RateFeatureFailure Rate
123.9%1027.1%1932.8%2844.6%
234.6%1130.2%2034.4%2941.3%
326.8%1238.0%2131.1%3032.6%
434.3%1335.5%2238.7%3133.8%
530.8%1432.3%2343.8%3227.2%
645.9%1525.1%2446.1%3333.9%
737.5%1626.2%2531.4%3421.4%
834.5%1734.3%2633.0%3524.7%
933.8%1837.1%2728.8%
Table 5. The statistical results of the correlation larger than 0.9.
Table 5. The statistical results of the correlation larger than 0.9.
13458910111931333435
11.000.900.830.72−0.090.010.600.590.01−0.06−0.09−0.07−0.04
3 1.000.760.73−0.340.330.520.510.080.02−0.010.000.01
4 1.000.94−0.230.060.610.61−0.11−0.16−0.21−0.19−0.16
5 1.00−0.06−0.140.500.50−0.16−0.22−0.25−0.22−0.19
8 1.00−0.94−0.16−0.160.140.110.180.180.22
9 1.000.130.130.000.05−0.02−0.03−0.08
10 1.001.000.420.420.330.340.33
11 1.000.430.430.340.350.35
19 1.000.910.880.680.84
31 1.000.970.810.84
33 1.000.730.81
34 1.000.99
35 1.00
Table 6. The redundant features based on the value of correlation in descending order.
Table 6. The redundant features based on the value of correlation in descending order.
The Features with the Correlation that Exceeds the Threshold 0.9The Correlation CoefficientThe Classification Error RateThe Feature that Need to be Removed
Feature 10 and 110.99860327.1% < 30.2%Feature 11
Feature 34 and 350.98586121.4% < 24.7%Feature 35
Feature 31 and 330.96871833.8% < 33.9%Feature 33
Feature 8 and 9−0.94469734.5% > 33.8%Feature 8
Feature 4 and 50.94461034.3% > 30.8%Feature 4
Feature 19 and 310.90731132.8% < 33.8%Feature 31
Feature 1 and 30.90106823.9% < 26.8%Feature 3
Table 7. The statistical results of pedestrian detection.
Table 7. The statistical results of pedestrian detection.
ClassifierFeature SetAccuracyTime/100 SamplesAUC
BPNN35 features81.6%6.19 ms0.898
18 features90.7%6.03 ms0.936
Adaboost35 features94.5%13.29 ms0.963
18 features93.2%8.68 ms0.952
Table 8. The results of pedestrian detection.
Table 8. The results of pedestrian detection.
MethodClassifier + Feature SetAccuracy
Segment classificationBPNN + 18 features80.7%
Adaboost + 18 features85.2%
Track classificationBPNN + 18 features93.7%
Adaboost + 18 features95.9%

Share and Cite

MDPI and ACS Style

Zhang, M.; Fu, R.; Cheng, W.; Wang, L.; Ma, Y. An Approach to Segment and Track-Based Pedestrian Detection from Four-Layer Laser Scanner Data. Sensors 2019, 19, 5450. https://doi.org/10.3390/s19245450

AMA Style

Zhang M, Fu R, Cheng W, Wang L, Ma Y. An Approach to Segment and Track-Based Pedestrian Detection from Four-Layer Laser Scanner Data. Sensors. 2019; 19(24):5450. https://doi.org/10.3390/s19245450

Chicago/Turabian Style

Zhang, Mingfang, Rui Fu, Wendong Cheng, Li Wang, and Yong Ma. 2019. "An Approach to Segment and Track-Based Pedestrian Detection from Four-Layer Laser Scanner Data" Sensors 19, no. 24: 5450. https://doi.org/10.3390/s19245450

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop