You are currently viewing a new version of our website. To view the old version click .
Remote Sensing
  • Article
  • Open Access

30 March 2017

A Density-Based Clustering Method for Urban Scene Mobile Laser Scanning Data Segmentation

,
,
,
and
1
School of Resource and Environment Sciences, Wuhan University, 129 Luoyu Road, Wuhan 430079, China
2
Beijing Institute of Architectural Design (Group) Co., Ltd, 62 Nanlishi Road, Xicheng District, Beijing 100045, China
3
Collaborative Innovation Centre of Geospatial Technology, Wuhan University, 129 Luoyu Road, Wuhan 430079, China
4
The Key Laboratory of GIS, Ministry of Education, China, Wuhan University, 129 Luoyu Road, Wuhan 430079, China

Abstract

The segmentation of urban scene mobile laser scanning (MLS) data into meaningful street objects is a great challenge due to the scene complexity of street environments, especially in the vicinity of street objects such as poles and trees. This paper proposes a three-stage method for the segmentation of urban MLS data at the object level. The original unorganized point cloud is first voxelized, and all information needed is stored in the voxels. These voxels are then classified as ground and non-ground voxels. In the second stage, the whole scene is segmented into clusters by applying a density-based clustering method based on two key parameters: local density and minimum distance. In the third stage, a merging step and a re-assignment processing step are applied to address the over-segmentation problem and noise points, respectively. We tested the effectiveness of the proposed methods on two urban MLS datasets. The overall accuracies of the segmentation results for the two test sites are 98.3% and 97%, thereby validating the effectiveness of the proposed method.

1. Introduction

With the developing ability to acquire high-quality point cloud data, mobile laser scanning systems have been widely utilized for various applications such as 3D city modeling [], road inventory studies [], safety control [], car navigation [], and forestry management []. Data from mobile laser scanning (MLS) systems, airborne laser scanning (ALS) systems, and terrestrial laser scanning (TLS) systems are being widely applied in urban scene analysis. However, MLS data are more suitable for urban scene information extraction for two reasons. First, compared to ALS data, MLS data are of higher density and contain more vertical information, factors that are of great importance in identifying detailed information from poles and buildings. Moreover, MLS data are more efficiently acquired than TLS data, as the latter are collected by manually positioned systems. Many researchers have studied the use of MLS data in urban scenes intensively, including in road [,] and road marking [,,] detection, building detection and reconstruction [,,], pole-like object detection [,,], tree detection and modeling [,,], and urban scene segmentation and classification [,,,,].
The segmentation of MLS point clouds into meaningful segments is a more difficult task than the extraction of a single class of street objects. Segmentation is a process partitioning point clouds into disconnected, salient segments and usually assumes that segments tend to represent individual objects or specific parts of objects []. Segmentation is a crucial and foundational step for further street object extraction and classification in urban scenes []. The major challenges for MLS data segmentation in urban scenes originate from three sources. The first challenge comes from the nature of point cloud data, which are characterized by their unorganized structure, uneven density, and huge data volume. Second, the purpose of segmentation is to find some universal criterion to segment MLS point cloud into discrete objects, although the objects often have different sizes and shapes. The last challenge is derived from the scene complexity of urban environments, where cars remain close to each other, trees and pole-like objects are tangled with each other, and various objects can lie under trees and poles.
The segmentation of ALS data has been intensively studied in the past decade [,,]. However, methods for the segmentation of ALS data are only partially applicable for MLS data due to differences in point densities, scanning patterns, and geometric characteristics []. Recently, to overcome the above-mentioned challenges, several methods have been proposed to segment MLS data. Most of these researchers have made great contributions to this area and could segment MLS data with high precision in many different scenes of urban areas. However, their methods still need further improvement to overcome under- and over-segmentation problems in complex scenes where objects are tangled with each other or where a huge variety of point densities exist.
We proposed a method that can effectively segment MLS point clouds in a complex scene with tangled objects and large data density variations. This paper is organized as follows. Related work on MLS data segmentation is presented in Section 2. The proposed method is described in detail in Section 3. The results of two experiments on urban scenes and some comments about them are presented in Section 4. The conclusions about the advantages and disadvantages of the proposed method and future work regarding our research are discussed in Section 5.

3. Methods

The proposed method attempts to segment MLS data from urban scenes into discernible and meaningful segments that can be applied for further street furniture extraction and classification. The method consists of three main phases (Figure 1):
Figure 1. The overall workflow of the proposed method.
  • Pre-processing: Original un-organized MLS data are cleaned and re-organized based on voxels; then, the whole scene is classified into ground and non-ground voxels.
  • Clustering: A density-based clustering method is utilized to segment the non-ground voxels into discrete clusters.
  • Post-processing: Voxels with cluster labels are back-projected points to merge clusters that belong to an individual street object accurately, and noise points generated in the clustering stage are re-assigned to the clusters.

3.1. Pre-Processing

This phrase attempts to provide clean and organized data for the following clustering algorithm, including the voxelization of points into voxels and the detection of ground voxels. Due to the complexity of the scanning environments in urban scenes, noise points that are isolated from the remaining point cloud are brought into the original MLS data. These noise points negatively affect the following algorithm; thus, they need to be detected and removed from the original MLS data. Isolated points are detected and removed using the connected component analysis method, and small clusters are considered as noise points and filtered out from the original MLS data.

3.1.1. Voxelization

Original point clouds are both large in volume and unorganized in structure; direct operations on the original MLS data will be highly time and memory consuming. Therefore, after filtering out isolated noise points, we implement a voxelization method similar to [], which regularly re-organizes and condenses the original data into a 3D space. In this context, a voxel is a cube that records three classes of information: voxel location, voxel index, and the number of points in the voxel. A voxel location is represented by three numbers ( n r o w , n c o l u m n , n h e i g h t ) , which represent the location of the voxel relative to the minimum x, y, and z position of the original point cloud. The location of the voxel can be computed using the following equation:
n r o w = i n t e g e r ( x x m i n V S )
where x m i n denotes the minimum x of all points in the point cloud and V S is the voxel size. n c o l u m n and n h e i g h t are calculated in a similar way.
After the processing step of the voxelization, all needed information for steps before the merging processing step are stored in the voxels, and all steps before merging are applied to the voxel set.

3.1.2. Ground Detection

A whole urban scene point cloud can be roughly classified into off-ground points, on-ground points, and ground points []. The segmentation targets in this paper are street furniture that stands on or close to the ground. To select these targets, the ground voxels need to first be detected. Numerous researchers have focused their effort on the detection or even reconstruction of roads from point cloud data [,,,,,,,,]. In addition, roads can sometimes play the role of ground to provide the relative position of the street furniture to the ground in situations where the ground is mainly even vertically along the trajectory. Nevertheless, in this context, complex situations of wide and uneven streets are considered. Moreover, in our next processing step for calculating parameters of the clustering algorithm, the objects’ distances to the ground rather than to the road need to be calculated. Therefore, we introduced a new ground detection method applied to the voxel set generated in the previous step.
The ground is assumed to be relatively low in a local area and is erected low vertically compared to the on-ground objects. Therefore, the ground can be separated from the whole scene by analyzing the relative height and the vertically continuous height ( H v ) of each horizontally lowest voxel (HLV) in the whole voxel set. The height of a voxel is represented by n h e i g h t , as depicted in Equation (1). Voxels that satisfy the conditions in Equation (2) are recognized as ground voxels:
H v < 1 / V S   & &   H r < 0.5 / V S
where H v is the vertically continuous height of an HLV, H r is the relative height of an HLV to the lowest voxel in its neighborhood, and V S is the voxel size. The vertical continuity analysis algorithm for calculating H v is depicted in detail in a previous study (Li et al., 2016). Utilizing the proposed algorithm, curbs and low-elevation bushes are also classified as ground and are thus filtered out before clustering.

3.2. Clustering

After ground voxels are detected from the original voxel set, the remaining voxels need to be segmented for the further extraction or classification of street objects. To this end, we proposed a clustering-based method for segmentation. Inspired by the algorithm implementing clustering via fast search and finding density peaks [], we also make two assumptions about cluster centers: (1) the center of a cluster is surrounded by voxels with lower local densities; and (2) the cluster centers are relatively far away from other cluster centers compared to its local neighboring voxels. Detailed information about the clustering method is descripted in the two sections below.

3.2.1. Generation of Cluster Centers

Assume that the on-ground voxel set for segmentation is V s = { v i } i = 1 N , where N is the number of voxels in V s . As presented in the assumptions before, two parameters play key roles in the definition of cluster centers, ρ i and δ i , which represent the local density parameter and the minimum distance parameter of a voxel v i , respectively.
Local density: The local density parameter ρ i mainly indicates the vertically continuous height H v i of a voxel at one horizontal location v i ( x i ,   y i ) , and it is also affected by two other factors: the vertical position of the voxel h i , and the number of points in the voxel p i . For every voxel v i , the density parameter ρ i can be formulated as
ρ i = { H v i h i H i + p i p m a x , d g r o u n d i < D t ( H v i h i H i + p i p m a x ) / d g r o u n d i ,   d g r o u n d i D t   .
where d g r o u n d i depicts voxel v i ’s vertical distance to the ground specified by Equation (4). In Equation (4), v j is the horizontally closest voxel to v i in the ground voxel set. D t is a ground distance threshold value based on real situations, and p m a x denotes the maximum number of points in one voxel.
d g r o u n d i = n h e i g h t i n h e i g h t j
Minimum distance: The minimum distance parameter δ i indicates the distance between the current voxel v i and the closest voxel that has a higher density as specified by Equation (3). d i j indicates a customized distance between voxels v i and v j . The distance can be calculated according to Equation (5) as follows:
d i j = { d i j E , L i = L j D n e i g h b o r , L i L j
where L i indicates the label of voxel v i specified by the three-dimensional seed filling algorithm, D n e i g h b o r indicates the radius for searching neighbors, and d i j E is the Euclidian distance between the voxels v i and v j .
For every voxel v i , the distance parameter δ i can be measured by the following formula:
δ i = { min j I S i { d i j } , I S i D n e i g h b o r , I S i =
where I s i is a voxel set, whose elements have a higher density than voxel v i and lie within a neighbor distance D n e i g h b o r to voxel v i . I s i can be measured by the following formula:
I s i = { v k I S : ρ k > ρ i & & d k i < D n e i g h b o r   }
Cluster center: After the local density parameter and minimum distance parameter for every voxel have been calculated, the cluster centers can be determined. Based on the two above-mentioned assumptions, the cluster centers should have both high local density value and a minimum distance value. Therefore, voxels with both a higher value than the local density threshold value ρ t and a minimum distance threshold value δ t are considered to be cluster centers. Cluster centers in the proposed method have practical meaning; they normally depict the highest part of street objects. The local density parameter generally indicates the height of an object, and the minimum distance parameter indicates the distance between two objects in certain ways. Assume that C = { c i } i = 1 N represents the label of each voxel and that S = { s j } j = 1 N b represents the index of the cluster centers. The label of each voxel is determined based on the formula below:
c i = { k ,   i S ( i = s k ) 1 , i S
An example for describing how to determine the cluster centers of a voxel set is presented in Figure 2. To specify our method more directly, we simplified the situation and projected a real scene onto a two-dimensional plane. One green rectangle in the figure represents one voxel, and the numbers in the voxel depict the sequence of their local density, where smaller numbers correspond to higher local densities. From the figure, it can be concluded that the two cluster centers at the bottom part of two street objects can easily be found, as they have both high local density and minimum distance values.
Figure 2. An example of selecting cluster centers: (a) voxels numbered by descending local density values; (b) local density and minimum distance value distribution of the voxels from (a).

3.2.2. Clustering

Based on the assumptions about cluster centers, they can be determined by choosing voxels that have both higher local density and minimum distance values. Then, the label of each voxel can be obtained in the sequence of local densities in descending order. Every non-center voxel’s label is determined by the neighboring voxel, which is the closest voxel that has a higher local density than the current voxel within a neighboring distance. Assume that the voxel set sorted with the local density by descending order is { m i } i = 1 N . The neighboring voxel index n m j of each voxel v i can be specified by the following formula:
n m j = { argmin j < i { d m i m j } , i > 1   & &   δ i D n e i g h b o r 0 ,    i = 1   | |   δ i = D n e i g h b o r
where δ i is the minimum distance parameter defined in Equation (6).
When the neighbor voxel of each voxel in the voxel set is configured, the clustering procedure is performed based on the sequence of local densities in descending order, as described in detail in Algorithm 1.
Algorithm 1: Clustering
Input:
   V s : voxel set for clustering
Parameters:
  N: total amount of voxel
D n e i g h b o r : radius for searching closest neighbors
Start:
(1) Calculate R = { ρ i } i = 1 N based on Equation (3).
(2) Sort R by descending order SR = { m i } i = 1 N .
(3) Calculate D = { δ i } i = 1 N from Equation (6).
(4) Calculate CN = { n i } i = 1 N from Equation (9).
(5) Initialize the label of each voxel from Equation (8).
(6) for each voxel v i in SR repeat:
(7)   if c m i = 1 :
(8)      c m i = c n m i
End
Output:
  C: cluster labels of V s
After the clustering of the voxel set, the most narrowly expanded objects, such as most trees, light poles, and cars, are well segmented. However, some widely expanded objects are over-segmented, which requires further processing. Moreover, based on the clustering algorithm, some voxels that have high minimum distance values but low local density values are not labeled. These voxels are regarded as halo voxels and need to be processed in the following re-assignment step.

3.3. Post-Processing

Once the clustering processing step is completed, the whole voxel set is segmented into clusters with different labels. However, the size of the cluster is constrained to a fixed size around the cluster center due to the limitation of the radius in finding the neighboring voxels, which leads to over-segmentation and halo voxel problems for large street objects. Therefore, further post-processing steps are needed to address these problems. First, a merging step applied to the back-projected point cloud is used to merge the clusters of large objects. Then, we present a re-assignment processing step to address those halo voxels generated in the previous clustering stage, in which some of these voxels are filtered out and others are merged with the nearest segments.

3.3.1. Merging of Clusters

Most trees and pole-like objects are often differentiable after the clustering stage because the pole-part of these objects can always play the role of a cluster center and because the spaces between different pole-parts are recognizable. However, large street objects are often over-segmented into various parts. Consequently, a merging step is introduced in this method to address the over-segmentation problem. To improve the accuracy, the voxels are first back-projected to points. The following processing steps, including the merging and refinement step, are both applied to the point cloud instead of the voxel set. As presented in the voxelization step, a one-to-one correspondence is built between each point and each voxel; hence, we can obtain a point cloud with corresponding labels through the voxel index stored at each point and each voxel.
In the merging step, a new merging algorithm is incorporated to merge neighboring clusters via region growing. The merging criteria are the connectivity and the curvature similarity of two clusters at their common borders. The connectivity of two clusters C L i and C L j is measured by the distance between them, which is the distance between the closest points in two clusters defined by the following formula:
D C i j = m i n ( d ( p m i , p n j ) )
where p m i and p n j represent points in the clusters C L i and C L j , respectively, and d ( p m i , p n j ) denotes the Euclidean distance between the two points. C L i and C L j can be recognized as neighboring clusters (NCs) only when D C i j is smaller than the threshold value of 0.5 m.
Apart from connectivity, the curvature similarity parameter measures the geometric shape similarity between two clusters. Previous methods often measure this similarity based on the curvature, normal vector, or principal direction similarity directly on the clusters (Zhou et al., 2012; Yang and Zhen, 2013). Nevertheless, to improve the segmentation accuracy, we introduced a method to measure the curvature similarity applied to pair points located at the common borders. Pair points have two points from two NCs, and the distance between them is lower than a threshold value. Assume that C L i and C L j are two NCs. The contained pair point set (PPS) can be defined as follows:
P P S i j = { ( p m i , p n j ) : d ( p m i , p n j ) < d t h , p m i C L i , p n j C L j }
where p m i is a point in C L i and p n j is a point in C L j . When the distance between them is less than d t h (0.5 m), ( p m i , p n j ) is considered to be an element of the PPS.
The curvature similarity of a point can be a measured as change rate of the normal vector which can be defined by the following formula:
C = e 3 e 1 + e 2 + e 3
where e 1 ,   e 2 , and e 3 are the eigenvalues of a point in descending order, which is calculated through Principal Component Analysis (PCA). In addition, the curvature similarity of two clusters can be measured as follows:
C N S i j = 1 2 k 1 k ( C a i j + C b i j )
where ( p a ,   p b ) is the pair point element in the pair point set P P S i j , and C a i j , C b i j are the corresponding curvatures of the pair point measured based on Equation (12).
Then, the rules for merging two clusters are as follows:
D C i j < d t h & & C N S i j < C T
where d t h is a threshold value already defined in Equation (11) and C T is a threshold value found by analyzing thousands of sample data points. The merging processing step traverses all the clusters generated in the previous step and generates new segments that correspond to street furniture at the object level by applying region growing.

3.3.2. Re-Assignment

The clustering stage generates halo voxels that have high minimum distance values but low local density values. These voxels often correspond to the borders of large trees that are far from the pole part of the tree and extruded parts or other parts of buildings that are not fully scanned by the laser beams. They are all not labeled in the clustering step, and they are also back-projected to points in the merging step. Therefore, we can process these halo points independently after the merging step. All these halo points are clustered using the connected component analysis algorithm, and then, the resulting clusters are merged with the closest segments generated in the merging step. However, those points that are far away from all segments are regarded as noise points and are filtered out.

4. Experiments

Two test sites from Wuhan city were chosen to test the effectiveness of our method. Test site 1 (TS-1) was acquired in the suburb area of Wuhan city, which has tangled street trees and poles in wide streets with high density variations. Test site 2 (TS-2) was located at Optical Valley, with substantially more street object categories; Optical Valley is a typical urban area in Wuhan City. Detailed information about the two test sites can be found in Table 1.
Table 1. Description of the test sites.

4.1. Voxelization and Ground Detection Results

According to the proposed method, the voxel size should be first decided prior to performing voxelization and then ground detection. The density of the mobile laser scanning data in each test site varies substantially both in between the test sites and at each test site. The voxel size should be configured based on the overall consideration of the above-mentioned criterion, the density of the test site, and the following voxel continuity judgment step of objects that are far away from the trajectory. The average point span for distant objects in the test sites is greater than approximately 0.15 m. To guarantee that distant objects have vertically continuous voxels, the voxel size was configured as twice the average point span: 0.3 m. The numbers of voxels after the voxelization of TS-1 and TS-2 are 170,864 and 157,674, respectively, with compression rates (1—number of voxels containing data/number of original points) of 97.5% and 92.1%, which will reduce the computation cost greatly compared to direct operation on point clouds. After voxelization, the ground can be detected based on Equation (2). The results for each test site after back-projection from the voxels to the points are presented in Figure 3, where the orange points are ground points.
Figure 3. Ground detection results after back-projecting to point clouds: (a) the ground detection result of test site 1 (TS-1); (b) the ground detection result of test site 2 (TS-2).

4.2. Clustering Results

After filtering out the ground voxels, the clustering-based segmentation processing step is then applied to the remaining voxel set, which produces cluster centers and subsequently segments from these cluster centers. The cluster centers are determined based on two key parameters, the local density and minimum distance, which are defined in Equations (3) and (6) in Section 3.2.1. The parameter configuration for the generation of cluster centers can be found in Table 2.
Table 2. Parameter configurations.
It is clearly impracticable to detect every small object in a complicated urban scene; hence, only the objects that fully satisfy the following criteria are considered in our experiment: (i) a height of greater than 1.2 m; (ii) a separable Euclidean distance of at least 0.9 m between the cluster centers of street objects and (iii) a distance of at most 1.5 m to the ground.
The values of ρ t and δ t are configured based on the first two assumptions about the targeted street objects that we focus on. The parameter value configuration of D t should guarantee that the off-ground parts of street objects, such as tree crowns or the board parts of traffic signs, will not form cluster centers in the clustering stage. Then, these objects will just have one cluster center and will not be segmented into various parts. Therefore, the value will be configured based on real situations, thus guaranteeing that it is smaller than the average distance between the off-ground parts of street objects and the ground. D n e i g h b o r is a parameter that is used to specify the radius that a voxel uses to search for neighboring voxels to calculate the minimum distance values. The value of D n e i g h b o r is configured as 3.9 m to ensure that most trees are not over-segmented, which results in greater merging and re-assignment work during the post-processing step. The parameter configuration for other datasets can be decided by concerned targets in real situations and can use the assumptions about our targeted street furniture as a reference.
The cluster center generation results for a typical scene are presented in Figure 4. A point in the figure corresponds to a voxel and is located at the center of the voxel. The cluster center generation results of the selected part of the test area are depicted in red in Figure 4h. From this figure, it can be concluded that all street objects can be located based on the cluster centers labeled in the figure.
Figure 4. Cluster center generation results in a typical scene: (a) original voxel set colored by the Z coordinate; (b) selected voxel set in black rectangle from (a) colored by the Z coordinate; (c) local density value distribution of the original voxel set; (d) local density value distribution of the selected voxel set; (e) minimum distance value distribution of the original voxel set; (f) minimum distance value distribution of the selected voxel set; (g) cluster center (colored in red) generation results of the original voxel set; (h) cluster center (colored in red) generation results of the selected voxel set.
After the cluster centers are determined, the clustering algorithm is applied to generate segments based on Algorithm 1. The clustering results for each test site are depicted in Figure 5. The numbers of segments generated in each test site are 311 and 544. It can be seen in the figure that in this stage, a majority of trees, poles, cars, and many other street objects, except buildings and fences, are well segmented at the object level. The over-segmented buildings and fences need to be merged in the subsequent merging step. There are also some segments that are too far away from the ground. Thus, they are recognized as halo voxels and filtered out in this stage, and they will be re-processed at the refinement stage in the post-processing stage.
Figure 5. Clustering results after back-projection to points for the test sites: (a) an overall scene from TS-1; (b) typical tangled trees and poles that are well-segmented; (c) the overall scene of TS-2; (d) buildings that are over-segmented.

4.3. Merging and Re-Assignment Results

After the merging process, 202 and 380 segments were generated in TS-1 and TS-2, respectively. Figure 6 depicts the results after merging for the two test sites. We can conclude that fences were successfully merged to form meaningful street objects after our merging method was applied. Moreover, some building failed to merge, as they were only partially reached by the laser beams because of the occlusions caused by the trees in front of them and the large space between the buildings and the laser scanner. However, it is also meaningful to segment buildings to this extent because they can also be detected in the subsequent extraction or classification stage if the overall geometry, such as the width of the segment, is not considered.
Figure 6. Results after merging: (a) the merging result from TS-1; (b) the merging result of selected area in (a); (c) the merging result of TS-2; (d) the merging result of selected area in (c).
The re-assignment processing step is then performed after merging to address these halo points. These noise points are first clustered and then assigned to the segments generated in the merging step according to the distance between them (Figure 7). It can be concluded that the noise clusters mainly originate from three sources. One source is the high part of buildings that cannot be reached completely by the laser beams. The second source is trees that are located far away from the trajectory, which leads to a low reach rate of the laser beams and occlusion by the trees in front of them. The last source comes from large road signs, which largely extend in the horizontal direction.
Figure 7. Results of the re-assignment step (red colored points in (a,c) represent the halo points that were re-assigned to (b,d) individually).

4.4. Performance Analysis of the Final Results

Most previous researchers evaluated segmentation results by analyzing the subsequent processing step after segmentation such as in the detection of street objects [,,] or classification based on segmentation results [,,,]. Few researchers have focused on the direct evaluation of segmentation results for two reasons. The first reason is that there are no standard ground truth data for the experimental data, as in the image segmentation research field, and manual labeling is rather time consuming. The second reason is that urban scenes sometimes are so complex that the street object boundaries are impossible to recognize manually. We attempt to directly evaluate the segmentation results at the object level by introducing two simple evaluation metrics: the under-segmentation rate (USR) and the over-segmentation rate (OSR). USR represents the proportion of under-segmented objects in the target segmented objects in the scene, while OSR denotes the ratio of over-segmented objects in all target objects. In addition, we also calculate the overall accuracy (OA) taking these two metrics into account.
Table 3 lists the results for the two test sites with our evaluation metrics for each type of object as well as the overall scene. We can conclude that these criteria can successfully reveal the quality of the results in a direct manner, as they concentrate on evaluating the results at the object level.
Table 3. Performance analysis for the segmentation results of the two test sites.
It can be seen that the proposed method can achieve high accuracy in segmenting the data of both test sites (OA of 98.3% and 97%), with tolerable minor errors. First, the method can not only perform well in simple situations with differentiable space between objects but can also obtain high accuracy in complex situations where trees and poles are tangled together (Figure 8a,b). Second, TS-1 had two lines of trees along the street with large density variations; our method can overcome the uneven density problem in TS-1 and can also segment the trees laying behind with first row of trees and poles with a lower density (Figure 8c). Moreover, the method is also robust in terms of addressing buildings with complicated structures in TS-2 (Figure 8d). Although some low density and occluded parts of the building were recognized as noise points at the clustering stage, they are re-assigned to the building successfully at the final stage of the method. The segmentation errors in TS-1 mainly result from under-segmentation problems, mainly caused by the nesting of objects. For example, two under-segmented trees are tangled with traffic signs in TS-1, and the pole parts are nested within each other, which makes it difficult to differentiate them correctly (Figure 8e,f). One tree was over-segmented because one traffic sign with a low height was laid right down upon it (Figure 8g). Two similar under-segmentation results also existed in TS-2, namely, those present in Figure 8e,f. One tree was over-segmented for a reason similar to that in Figure 8g. The major over-segmentation originates from the buildings in TS-2, where three buildings are over-segmented for two reasons. First, some building segments are far away from the segments in the same building because of the absence of laser beams in certain parts of the segments due to occlusions (Figure 8h). The other source of errors is the irregular distribution of the points at the common border of two over-segmented building parts, which lead to the unsuccessful merging of two segments in one building (Figure 8i). However, most of the buildings in the test sites are well segmented if they are not occluded excessively. In addition, these over-segmentation errors will not inhibit the recognition of these segments as buildings as long as the sizes of these segments are not used as the segment features.
U S R = N u m b e r   o f   o b j e c t s   i n   s e g m e n t s   w i t h   m o r e   t h a n   o n e   o b j e c t s   N u m b e r   o f   o b j e c t s
O S R = N u m b e r   o f   o b j e c t s   s e g m e n t e d   t o   m o r e   t h a n   o n e   s e g m e n t N u m b e r   o f   o b j e c t s
O A = 1 U S R + O S R 2
Figure 8. Typical scenes in the test sites: (a) tangled trees and lamps; (b)another scene of tangled trees and lamps; (c) objects of variable densities; (d) buildings with complicated structure; (e) nested trees and traffic signs; (f) nested trees and traffic signs; (g) over-segmented trees with a traffic sign standing below; (h) occluded buildings that are over-segmented; (i) over-segmented buildings because of irregular points distribution.

5. Conclusions

This paper proposes a density-based clustering approach to segment urban scene MLS data into objects. First, after filtering out the noisy points, the original point cloud dataset is voxelized, and the ground voxels are detected based on the assumptions that we make about them. In the second stage, the key clustering processing step is performed. The cluster centers are found based on two key parameters: local density and minimum distance. In addition, the labeling process is performed in descending order of each voxel’s local density value. In the final stage, a merging step is first applied to the back-projected point cloud to merge those segments generated in the second stage. Finally, a re-assignment step for processing the noise points is conducted to produce the final segmentation results.
The density-based clustering method proves to be robust in tangled situations; for example, it is efficient at differentiating individual trees when their branches are connecting with each other, which other methods may find difficult (Figure 8a,b). Besides, the clustering method with the new definition of local density makes it possible to segment out different objects in mobile laser scanning data with large density variations (Figure 8c). Moreover, the proposed method does not require any additional information except for the coordinates of the point clouds which makes it applicable for more datasets. The experimental results show that our method can effectively segment urban mobile laser scanning data in variable cases, with an overall accuracy of greater than 97% based on our proposed criteria. The results also indicate that the proposed method can perform well not only in simple situations where there is discernible space between the objects, but also in complicated scenes where trees and poles are tangled together. However, as presented in Section 4.4, the poles and trees that are too nested with each other cannot be segmented well; this requires further improvement in our future studies.

Acknowledgments

This study is funded by Scientific and Technological Leading Talent Fund of National Administration of Surveying, mapping and geo-information (2014), the Wuhan ‘Yellow Crane Excellence’ (Science and Technology) program (2014), and the Fundamental Research Funds for the Central Universities (2012205020211).

Author Contributions

All authors contributed to the design of the methodology and the validation of experimental exercise; You Li and Lin Li wrote the paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sampath, A.; Shan, J. Segmentation and reconstruction of polyhedral building roofs from aerial LIDAR point clouds. IEEE Trans. Geosci. Remote Sens. 2010, 48, 1554–1567. [Google Scholar] [CrossRef]
  2. Rodríguez-Cuenca, B.; García-Cortés, S.; Ordóñez, C.; Alonso, M.C. An approach to detect and delineate street curbs from MLS 3D point cloud data. Autom. Constr. 2015, 51, 103–112. [Google Scholar] [CrossRef]
  3. Pu, S.; Rutzinger, M.; Vosselman, G.; Oude Elberink, S. Recognizing basic structures from mobile laser scanning data for road inventory studies. ISPRS J. Photogramm. Remote Sens. 2011, 66, S28–S39. [Google Scholar] [CrossRef]
  4. Brenner, C. Global localization of vehicles using local pole patterns. In Pattern Recognition; Springer: Berlin/Heidelberg, Germany, 2009; pp. 61–70. [Google Scholar]
  5. Zhang, C.; Zhou, Y.; Qiu, F. Individual tree segmentation from lidar point clouds for urban forest inventory. Remote Sens. 2015, 7, 7892–7913. [Google Scholar] [CrossRef]
  6. Yang, B.; Fang, L.; Li, J. Semi-automated extraction and delineation of 3D roads of street scene from mobile laser scanning point clouds. ISPRS J. Photogramm. Remote Sens. 2013, 79, 80–93. [Google Scholar] [CrossRef]
  7. Chen, D.; He, X. Fast automatic three-dimensional road model reconstruction based on mobile laser scanning system. Opt. Int. J. Light Electron Opt. 2015, 126, 725–730. [Google Scholar] [CrossRef]
  8. Yang, B.; Fang, L.; Li, Q.; Li, J. Automated extraction of road markings from mobile LIDAR point clouds. Photogramm. Eng. Remote Sens. 2012, 78, 331–338. [Google Scholar] [CrossRef]
  9. Li, L.; Zhang, D.; Ying, S.; Li, Y. Recognition and reconstruction of zebra crossings on roads from mobile laser scanning data. ISPRS Int. J. Geo-Inf. 2016, 5, 125. [Google Scholar] [CrossRef]
  10. Guan, H.; Li, J.; Yu, Y.; Wang, C.; Chapman, M.; Yang, B. Using mobile laser scanning data for automated extraction of road markings. ISPRS J. Photogramm. Remote Sens. 2014, 87, 93–107. [Google Scholar] [CrossRef]
  11. Pu, S.; Vosselman, G. Knowledge based reconstruction of building models from terrestrial laser scanning data. ISPRS J. Photogramm. Remote Sens. 2009, 64, 575–584. [Google Scholar] [CrossRef]
  12. Jochem, A.; Höfle, B.; Rutzinger, M. Extraction of vertical walls from mobile laser scanning data for solar potential assessment. Remote Sens. 2011, 3, 650–667. [Google Scholar] [CrossRef]
  13. Vo, A.-V.; Truong-Hong, L.; Laefer, D.F.; Bertolotto, M. Octree-based region growing for point cloud segmentation. ISPRS J. Photogramm. Remote Sens. 2015, 104, 88–100. [Google Scholar] [CrossRef]
  14. Li, L.; Li, Y.; Li, D. A method based on an adaptive radius cylinder model for detecting pole-like objects in mobile laser scanning data. Remote Sens. Lett. 2016, 7, 249–258. [Google Scholar] [CrossRef]
  15. Cabo, C.; Ordóñez, C.; García-Cortés, S.; Martínez, J. An algorithm for automatic detection of pole-like street furniture objects from mobile laser scanner point clouds. ISPRS J. Photogramm. Remote Sens. 2014, 87, 47–56. [Google Scholar] [CrossRef]
  16. Li, L.; Li, D.; Zhu, H.; Li, Y. A dual growing method for the automatic extraction of individual trees from mobile laser scanning data. ISPRS J. Photogramm. Remote Sens. 2016, 120, 37–52. [Google Scholar] [CrossRef]
  17. Wu, B.; Yu, B.; Yue, W.; Shu, S.; Tan, W.; Hu, C.; Huang, Y.; Wu, J.; Liu, H. A voxel-based method for automated identification and morphological parameters estimation of individual street trees from mobile laser scanning data. Remote Sens. 2013, 5, 584–611. [Google Scholar] [CrossRef]
  18. Rutzinger, M.; Pratihast, A.; Oude Elberink, S.; Vosselman, G. Detection and modelling of 3D trees from mobile laser scanning data. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2010, 38, 520–525. [Google Scholar]
  19. Zhou, Y.; Yu, Y.; Lu, G.; Du, S. Super-segments based classification of 3d urban street scenes. Int. J. Adv. Robot. Syst. 2012. [Google Scholar] [CrossRef]
  20. Yang, B.; Zhen, D. A shape based segmentation method for mobile laser scanning point clouds. ISPRS J. Photogramm. Remote Sens. 2013, 81, 19. [Google Scholar] [CrossRef]
  21. Yang, B.; Dong, Z.; Zhao, G.; Dai, W. Hierarchical extraction of urban objects from mobile laser scanning data. ISPRS J. Photogramm. Remote Sens. 2015, 99, 45–57. [Google Scholar] [CrossRef]
  22. Vosselman, G. Point cloud segmentation for urban scene classification. ISPRS Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2013, 1, 257–262. [Google Scholar] [CrossRef]
  23. Aijazi, A.; Checchin, P.; Trassoudaine, L. Segmentation based classification of 3D urban point clouds: A super-voxel based approach with evaluation. Remote Sens. 2013, 5, 1624–1650. [Google Scholar] [CrossRef]
  24. Barnea, S.; Filin, S. Segmentation of terrestrial laser scanning data using geometry and image information. ISPRS J. Photogramm. Remote Sens. 2013, 76, 33–48. [Google Scholar] [CrossRef]
  25. Filin, S.; Pfeifer, N. Segmentation of airborne laser scanning data using a slope adaptive neighborhood. ISPRS J. Photogramm. Remote Sens. 2006, 60, 71–80. [Google Scholar] [CrossRef]
  26. Oude Elberink, S.; Kemboi, B. User-assisted object detection by segment based similarity measures in mobile laser scanner data. ISPRS Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2014. [Google Scholar] [CrossRef]
  27. Zhou, Y.; Wang, D.; Xie, X.; Ren, Y.; Li, G.; Deng, Y.; Wang, Z. A fast and accurate segmentation method for ordered lidar point cloud of large-scale scenes. IEEE Geosci. Remote Sens. Lett. 2014, 11, 1981–1985. [Google Scholar] [CrossRef]
  28. El-Halawanya, S.I.; Lichtia, D.D. Detecting road poles from mobile terrestrial laser scanning data. GISci. Remote Sens. 2013, 50, 704–722. [Google Scholar]
  29. Golovinskiy, A.; Kim, V.G.; Funkhouser, T. Shape-based recognition of 3d point clouds in urban environments. In Proceedings of the 2009 IEEE 12th International Conference on Computer Vision, Kyoto, Japan, 27 September–4 October 2009. [Google Scholar]
  30. Vosselman, G.; Gorte, B.; Sithole, G.; Rabbani, T. Recognising structure in laser scanner point clouds. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2004. [Google Scholar] [CrossRef]
  31. Rabbani, T.; van den Heuvel, F.; Vosselmann, G. Segmentation of point clouds using smoothness constraint. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2006, 36, 248–253. [Google Scholar]
  32. Rodríguez-Cuenca, B.; García-Cortés, S.; Ordóñez, C.; Alonso, M. Automatic detection and classification of pole-like objects in urban point cloud data using an anomaly detection algorithm. Remote Sens. 2015, 7, 12680–12703. [Google Scholar] [CrossRef]
  33. Demantké, J.; Mallet, C.; David, N.; Vallet, B. Dimensionality based scale selection in 3d lidar point clouds. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2011. [Google Scholar] [CrossRef]
  34. Toth, C.; Paska, E.; Brzezinska, D. Using road pavement markings as ground control for lidar data. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2008, 37, 189–195. [Google Scholar]
  35. Riveiro, B.; González-Jorge, H.; Martínez-Sánchez, J.; Díaz-Vilariño, L.; Arias, P. Automatic detection of zebra crossings from mobile lidar data. Opt. Laser Technol. 2015, 70, 63–70. [Google Scholar] [CrossRef]
  36. Kumar, P.; McElhinney, C.P.; Lewis, P.; McCarthy, T. Automated road markings extraction from mobile laser scanning data. Int. J. Appl. Earth Obs. Geoinf. 2014, 32, 125–137. [Google Scholar] [CrossRef]
  37. Chen, X.; Kohlmeyer, B.; Stroila, M.; Alwar, N.; Wang, R.; Bach, J. Next generation map making: Geo-referenced ground-level lidar point clouds for automatic retro-reflective road feature extraction. In Proceedings of the 17th ACM SIGSPATIAL International Conference on Advances in Geographic Information Systems, Seattle, WA, USA, 4–6 November 2009; pp. 488–491. [Google Scholar]
  38. Tighe, J.; Niethammer, M.; Lazebnik, S. Scene parsing with object instances and occlusion ordering. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 3748–3755. [Google Scholar]
  39. Golparvar-Fard, M.; Balali, V.; Garza, J.M.D.L. Segmentation and recognition of highway assets using image-based 3d point clouds and semantic texton forests. J. Comput. Civ. Eng. 2012, 29. [Google Scholar] [CrossRef]
  40. Balali, V.; Golparvar-Fard, M. Segmentation and recognition of roadway assets from car-mounted camera video streams using a scalable non-parametric image parsing method. Autom. Constr. 2015, 49, 27–39. [Google Scholar] [CrossRef]
  41. Balali, V.; Golparvarfard, M. Recognition and 3d localization of traffic signs via image-based point cloud models. In Proceedings of the 2015 International Workshop on Computing in Civil Engineering, Austin, TX, USA, 21–23 June 2015. [Google Scholar]
  42. Guan, H.; Li, J.; Yu, Y.; Chapman, M.; Wang, H.; Wang, C.; Zhai, R. Iterative tensor voting for pavement crack extraction using mobile laser scanning data. IEEE Trans. Geosci. Remote Sens. 2015, 53, 1527–1537. [Google Scholar] [CrossRef]
  43. Rodriguez, A.; Laio, A. Clustering by fast search and find of density peaks. Science 2014, 344, 1492–1496. [Google Scholar] [CrossRef] [PubMed]
  44. Serna, A.; Marcotegui, B. Detection, segmentation and classification of 3d urban objects using mathematical morphology and supervised learning. ISPRS J. Photogramm. Remote Sens. 2014, 93, 243–255. [Google Scholar] [CrossRef]

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.