You are currently viewing a new version of our website. To view the old version click .
ISPRS International Journal of Geo-Information
  • Article
  • Open Access

2 November 2020

A Novel Indoor Structure Extraction Based on Dense Point Cloud

,
and
1
College of Surveying and Geo-Informatics, Tongji University, Shanghai 200092, China
2
Key Laboratory of Urban Land Resources Monitoring and Simulation, Ministry of Natural Resources, Shenzhen 518034, China
*
Author to whom correspondence should be addressed.

Abstract

Herein, we propose a novel indoor structure extraction (ISE) method that can reconstruct an indoor planar structure with a feature structure map (FSM) and enable indoor robot navigation using a navigation structure map (NSM). To construct the FSM, we first propose a two-staged region growing algorithm to segment the planar feature and to obtain the original planar point cloud. Subsequently, we simplify the planar feature using quadtree segmentation based on cluster fusion. Finally, we perform simple triangulation in the interior and vertex-assignment triangulation in the boundary to accomplish feature reconstruction for the planar structure. The FSM is organized in the form of a mesh model. To construct the NSM, we first propose a novel ground extraction method based on indoor structure analysis under the Manhattan world assumption. It can accurately capture the ground plane in an indoor scene. Subsequently, we establish a passable area map (PAM) within different heights. Finally, a novel-form NSM is established using the original planar point cloud and the PAM. Experiments are performed using three public datasets and one self-collected dataset. The proposed plane segmentation approach is evaluated on two simulation datasets and achieves a recall of approximately 99%, which is 5% higher than that of the traditional plane segmentation method. Furthermore, the triangulation performance of our method compared with the traditional greedy projection triangulation show that our method performs better in terms of feature representation. The experimental results reveal that our ISE method is robust and effective for extracting indoor structures.

1. Introduction

Recently, the generation of three-dimensional (3D) models of the environment has garnered significant interest in various fields, e.g., building renovation [1], indoor navigation [2], architecture [3], and facility management [4,5]. With the availability of mature technology for data collection, rich environment information can be captured through many platforms such as light detection and ranging [6], RGB-D cameras [7], and stereo cameras [8]. Subsequently, efficient and robust structure extractions based on acquired data became a popular research topic. The purpose of indoor structure extraction (ISE) is to capture the salient structure and to provide a robust description of the interior layout. Traditional manual ISE is tedious and requires excessive manual intervention. Compared with the structure extraction of the outdoor environment [9,10], ISE is more challenging in terms of three aspects: it requires (a) a higher quantity of clutter, (b) a more complex geometry, and (c) a more uneven-distributed occlusions.
According to the data type, ISE methods are generally classified into two major categories: (1) ISE based on image data using formatted images and small data memory, subject to illumination changes, for which some studies based on image data have been performed [11,12,13], and (2) ISE based on point cloud data using an unstructured point cloud, favorable spatial topology, and a large memory footprint, for which studies have also been performed [14,15,16,17]. Furthermore, methods combining image and point clouds have been proposed [18,19,20], and video has been exploited [21,22,23].
Moreover, the resulting form of ISE can be classified into three types: feature structure map (FSM), semantic structure map (SSM), and navigation structure map (NSM). An FSM is constructed based on one or more types of indoor structures. Robust indoor structures can be represented accurately; however, topological associations of features are not considered. An SSM is generated using a semantic feature from an indoor scene. The major indoor layout or structure primitives can be reconstructed; however, a detailed description of some structures is insufficient and some semantic features are inessential for location and navigation. The NSM is mainly used to emphasize obstacle target detection, topological relations, and passable areas in indoor spaces; however, the indoor structure is not organized in a unified and reasonable form.
Herein, we propose a novel ISE method based on a point cloud. The result of our ISE comprises an FSM and an NSM. To construct the FSM, a novel two-stage region growing plane segmentation algorithm was used to extract planar features (the original planar point cloud) with better integrity. Subsequently, planar inliers were simplified using a quadtree-based segmentation algorithm. Finally, two triangulation strategies in the boundary and interior were used to reconstruct the planar structure. To construct the NSM, we used a novel NSM comprising a passable area map (PAM) within different heights and the original planar point cloud. Moreover, our NSM was constructed based on the Manhattan world assumption, which pertains to the statics of the city and indoor scenes. It states that an indoor scene is built on a Cartesian grid where surpatches in the world are aligned with one of three dominant directions. Our main contributions are four-fold: (1) The proposed two-stage region growing plane segmentation algorithm improves the integrity of the plane and the robustness of the FSM. (2) The quadtree simplification method can significantly reduce data memory, and the proposed triangulation strategy can emphasize the connection between the boundary and interior structure. (3) For NSM construction, our method can capture the ground well from indoor point clouds without additional prior constraints. (4) The proposed novel NSM form not only can indicate a passable area for path planning but also can serve as a reference feature map to improve the efficiency and robustness of indoor robot navigation.
The remainder of the paper is organized as follows. Related studies are reviewed in Section 2. Our ISE method is introduced in detail in Section 3. The experimental details are presented in Section 4. The experimental results are discussed and analyzed in Section 5. The conclusions are provided in Section 6.

3. Method

A flowchart of the proposed ISE method is shown in Figure 1. The input is a raw dense non-structured point cloud, and the output comprises an FSM and an NSM. The processing pipeline comprises four stages:
Figure 1. The overall processing pipeline of the proposed method.
  • Preprocessing: the raw point cloud data are processed to eliminate invalid and noise points. Subsequently, the data volume is reduced significantly via downsampling.
  • Plane segmentation: The planar structure is captured through a hybrid two-staged region-growing algorithm. Subsequently, the preprocessing point cloud is classified into the original plane and left point clouds.
  • FSM construction: After plane segmentation, each planar point cloud is classified into inliers and boundaries. Subsequently, different triangulation strategies are adopted in the two parts. Finally, an FSM (e.g., a planar mesh model) based on a planar structure is generated.
  • NSM construction: A novel ground plane is proposed based on indoor structure analysis under the Manhattan world assumption. Subsequently, an obstacle map within different heights and initial passable areas are constructed. Finally, a novel NSM comprising an original planar point cloud and a PAM is generated.

3.1. Preprocessing

The preprocessing comprises three steps: invalid point removal, statistical filtering, and voxel filtering. First, the invalid data type, “NAN”, is removed in the raw point cloud and M a p I is obtained. Subsequently, some error and noise points generated during scanning or stitching are eliminated via statistical filtering using Equation (1), and M a p S is acquired. Finally, voxel filtering with voxel size ( r r r ), a downsampling method, is performed to leverage the data memory. In addition, the sampling point cloud data M a p V is preserved.
{ p t n o i s e   |   p t   M a p I ,   n a d ( p t ) > ε d ¯   } ,
where n a d ( p t ) is the distance average between p t and n k neighbor points in M a p I , ε is a scale coefficient ( ε = 3 ), and d ¯ is the average distance of all points and their neighbor points in M a p I .

3.2. Two-Staged Region Growing Plane Segmentation

In indoor scenes, piecewise planar structures overwhelm other surface structures. Therefore, we reconstructed the indoor structure based on a planar feature. Two prominent approaches exist for plane segmentation: model fitting and region growing. The model fitting method is implemented on a predefined mathematical model. Two widely used models are random sample consensus (RANSAC) [59] and the Hough transform [60]. In the model fitting method, the spatial topology cannot be considered in a 3D space; hence, discontinuities may occur in the planar feature. The existing region growing method is sensitive to the computation accuracy of the growing conditions, which may impair the integrity of the planar feature. Therefore, we proposed a hybrid two-staged region growing plane segmentation to manage the boundary structure details more effectively. Our method comprises point-based region growing (stage 1) and plane-based region growing (stage 2).

3.2.1. Point-Based Region Growing

In Algorithm 1, the point-based region growing is illustrated. The algorithm uses M a p V as input and generates a preliminarily segmented plane P l A . The growing conditions, including the normal angle requirement and point-to-plane distance constraint, are adopted. In the growing analysis, the neighbor points with qualified normal are preserved in candidate seeds. In Equation (2), the point-to-plane distance is further calculated for points with unqualified normal. Points with a point-to-plane distance less than the distance threshold δ 1 will be preserved in the growing planar segment but not for the candidate seeds. The point cloud M a p V is partitioned into the preliminary plane P l A and the remaining points   R p after the point-based region grows.
d p t p = | p s , n t   +   d t | ( n t ) x 2 + ( n t ) y 2 +   ( n t ) z 2 ,
where p s is the query point, ( n t , d t ) is the parameter of the plane equation, and n t is the normal to the plane.
Algorithm 1 Point-based region growing (Stage 1)
1: Input: Point cloud   M a p V   { p t i |   p t i   M a p V } , neighbor finding function Ω ( . ) , angle threshold
2: θ 1 distance threshold δ 1 .
3: Output: Preliminary segmented plane point cloud { P l A }
4: Initialize: Plane   point   cloud   label   i d p l a n e 0 , point   neighbors   { N b d i }   { } , point normals
5: { N m l i } { } , point curvature { Cv i }   { } , region label { R l } { R l i   | R l i =   1 ,   i =   0 , 1
6: size ( M a p ) }, { N b d i }   Ω   ( p t i ) , { N m l i } PCA ( N b d i ), Cv i EVD ( N b d i ), point order
7: { Po } sort points according to curvature in ascending order.
8: Candidate seeds   { S c A } { } , insert the minimum curvature point in { P o } to { S c A }
9: if there is still R l i equals to −1
10: while { S c A } is not empty do
11:   current seed C s   the first element in { S c A } , remove C s from { S c A }
12:   for idx     { N b d i }   of   C s  
13:    angle compute angle between   { N m l i }   of   i d x   and   { N m l i }   of   C s
14:    if angle < θ 1
15:     insert idx to { S c A } ,     R l i d x i d p l a n e
16:    end if
17:    else
18:     dist compute distance between p t i d x and the fitting local plane of C s
19:     if dist   < δ 1
20:       R l i d x i d p l a n e
21:     end if
22:     else
23:       R l i d x   −2
24:     end else
25:    end else
26:   end for
27:  end while
28:   i d p l a n e increases by 1
29: end if
30: if R l i > −1
31:  Assemble p t i to P l i A according to R l i
32: end if
Where PCA is principal component analysis, and EVD is eigenvalue decomposition.

3.2.2. Plane-Based Region Growing

After the point-based region growing, some plane points, R p , such as the boundary points, remained unclassified owing to the limitation of the growing condition. Plane-based region growing was adopted to further process unclassified plane points.
(1)
Remaining point updates
As shown in Equation (3), we adopted a point quantity threshold ζ 1 to filter small planes P l t i and to scatter them back into R p . Subsequently, the updated remaining point cloud R p u is obtained.
R p u = R p +   P l t i ,   P l t i =   { P l i A | n ( P l i A ) <   ζ 1 } ,
where n ( P l i A ) denotes the point quantity of plane point cloud P l i A .
(2)
Plane Growing
Neighbor points N p i B for all points of P l i A are established in R p u , and repetitive points in N p i B are removed. A growing process similar to stage 1 was performed on N p i B and R p u . However, only the distance growing condition was considered. In Equation (4), each plane point cloud P l i B comprises qualified unclassified points P t q u a and the corresponding preliminary plane P l i A after the plane-based region grows.
P l i B = P l i A +   P t q u a ,   P t q u a =   { p t   | d ( p t ,   P l i A ) <   δ 2 ,   p t     N p i B } , i   =   0 ,   1 ,   2 ,   ,   n p ( P l i A ) ,
where d ( p t ,   P l i A ) is the distance from point p t to plane P l i A , δ 2 is the distance threshold, and np( P l i B ) is the plane quantity of P l i A .
Original planar point clouds P l B and updated remaining point clouds R p u were obtained after growing the plane-based region. Some of the plane segmentation results of our proposed two-stage region growing method are shown in Figure 2. The red points represent the remaining unclassified points, whereas the points with other colors represent different planes. As shown in Figure 2, a few red unclassified points appeared around the intersection of the yellow and gray walls. Two main aspects are involved in handling the boundary information: (1) In point-based region growing, the points around the boundary are classified into plane, but not the candidate seeds leveraging on the distance condition. This avoids early ceasing before reaching the boundary vicinity; (2) in plane-based region growing, boundary points in R p are further classified into planes based on the distance. This further reduces the quantity of unclassified points around the vicinity of the boundary.
Figure 2. Segmentation plane point cloud.

3.3. Plane Simplification

Although most planar structures have been captured using our plane segmentation method, considerable redundancy still existed. We adopted a quadtree segmentation based on cluster fusion to simplify the planar feature. The advantage of this method is that the image can be efficiently partitioned into regular, square clustering regions. The plane simplification pipeline is shown in Figure 3. In the algorithm, different strategies were adopted in the boundary and interior of the plane. The boundary points were maintained to guarantee the overall distribution of the planar structure. Some new points were generated in the interior to reconstruct the characteristics of the planar structure.
Figure 3. The pipeline of plane simplification.

3.3.1. Orthographic Projection

For each extracted plane point cloud P l i B , the plane normal N o i ( ω x , ω y , ω z ) was estimated based on the least square. The rotation matrix R o t in Equation (5) was obtained; then, the direction-corrected planar point cloud P l i d c was acquired by rotating P l i B with R o t . Subsequently, we obtained the projected point cloud P l i p r in Equation (6).
R o t = I + Κ   sin θ +   Κ 2 ( 1 cos θ ) ,   K = ( 0 ω z ω y ω z 0 ω x ω y ω x 0 )
P l i p r = { p   |   p . x   = q . x ,   p . y = q . y , p . z = 0 ,   q     P l i d c } ,
where θ is the angle between N o i and Z (0, 0, 1).

3.3.2. Image Generation

The boundary point cloud   B d of P l i p r was extracted using the alpha shape boundary method [61]. Subsequently, using Equation (7), we mapped P l i p r to a grayscale image I g i using function Λ . Next, we set the height and width of I g i in Equation (8) to be equivalent to prepare for the subsequent segmentation.
{   p o i n t   c l o u d     Λ     I g ,   I g   Λ p o i n t   c l o u d Λ :   r =   y m a x   y r e s   ,   c =   x     x m i n r e s    
w = h = { 2 s   |   2 s > max ( b r w ,   b r h ) >   2 s 1 ,   s N } ,
where x and y are the coordinates X and Y of each point in P l i p r , respectively; r and c are the row and column of each pixel in I g , respectively; r e s is the image pixel resolution; b r w and b r h are the width and height of the minimum bounding rectangle of P l i p r , respectively; and w and h are the width and height of I g , respectively.

3.3.3. Quadtree Segmentation

The pixels in I g i are classified into exterior ( E p ) , boundary ( B p ), and interior ( I p ). In Figure 4, the principle of quadtree segmentation is illustrated, where I g i is partitioned into four square regions each time. The segmentation process does not terminate until all elements are the same in each partitioned region. Subsequently, all B p and interior squares ( I s ) that contain I p are preserved. The four corners in   B p and I s are converted into point clouds B p p i S and I s p i S through function Λ . Finally, for each P l i B , a simplified plane point cloud P l i s comprising B p p i S and I s p i S is obtained.
Figure 4. The quadtree segmentation based on cluster fusion.
The simplified planar point cloud is shown in Figure 5. The result (Figure 5b) shows the boundary points accurately. As shown in Figure 5c, the grayscale values for E p , B p , and I p in I g i were set to 0, 100, and 240, respectively, for distinction. For the planar simplified result (Figure 5d), the outline of the planar segments was preserved, and the interior redundant points were removed.
Figure 5. Simplification results of plane point cloud: (a) original plane point cloud, (b) boundary point cloud in green, (c) grayscale image, and (d) simplified plane point cloud ( P l i s ).

3.4. Feature Triangulation

After the planar simplification, we reconstructed the model representation of the planar feature using different triangulation strategies in the interior and boundary of each P l i s . In the interior, a simple triangulation was performed, in which two triangles were directly generated based on the four corners of I s p i S . At the boundary, a vertex-assignment triangulation (Figure 6) was performed, and four rules were proposed for the triangulation process.
Figure 6. Flow chart of vertex-assignment triangulation.

3.4.1. Boundary Assignment

The proposed boundary triangulation uses B p p i S and I s p i S as inputs and generates the feature triangulation. Figure 7 shows our proposed distance rule. Figure 7a shows two types of relative positions between the points and edges as well as the proposed distance rule for calculating the distance from a point to an edge. Figure 7b shows the calculation of the distances from b p p j ( b p p j     B p p i S ) to the four edges in i s p k ( i s p k   I s p i S ) . e m , i.e., the edge with the minimum distance to b p p j , was determined, and the corresponding minimum distance was regarded as the distance from b p p j to i s p k . Subsequently, b p p j was categorized into the collected point set P s of e m . In Figure 7b, d 1 is the minimum distance and e 1 is the e m . Finally, the P s for all edges of i s p k was obtained.
Figure 7. Distance rule: (a) two sorts of relative position between point and edge and (b) determination of the closest edge. p t refers to a b p p point. e 1 4 represent four edges of i s p k . d 1 4 represent the distance from a b p p point to four edge in i s p k .

3.4.2. Boundary Sorting

For the non-empty P s on one edge, the projection point ϑ p t of b p p j on the corresponding edge was calculated. Figure 8 shows the proposed sorting for sorting the boundary points in each P s . The sorting rule was based on the coordinates of ϑ p t . The x- or y-coordinates were utilized to sort the boundary points based on the direction of the edge. Y and X were selected for the vertical and horizontal directions, respectively.
Figure 8. Sorting rule: ϑ p t 1 4 are the projection points. p t 1 and p t 2 are sorted according to the x coordinate of ϑ p t 1 and ϑ p t 2 . p t 3 and p t 4 are sorted according to the y coordinate of ϑ p t 3 and ϑ p t 4 .

3.4.3. Triangle Connection

After boundary sorting, we proposed a connection rule to construct the triangle preliminarily, as shown in Figure 9. For every non-empty P s point on one edge, we connected triangles through two neighbor boundary points. Four types of triangle connections were summarized based on the relative position between the edge and two neighbor points. In Figure 9, p t 1 and p t 2 represent two b p p points. Type 1 (Figure 9a) involves two points outside the same side of the edge; type 2 (Figure 9b) involves one point inside the edge and another point outside the edge; type 3 (Figure 9c) involves two points inside the edge; and type 4 (Figure 9d) involves two points outside the different sides of the edge. In type 4, a midpoint between two boundary points is generated to assist the connection of triangles.
Figure 9. Connection rule: (a) type 1, (b) type 2, (c) type 3, and (d) type 4. p t a is the midpoint of two neighbor b p p points.

3.4.4. Triangle Adjustment

After the preliminary triangle connection, some irrational triangles that intersect with other squares must be modified. Therefore, an adjustment rule was proposed to modify the inappropriate triangles, as shown in Figure 10. Figure 10a shows triangle ( l 1 , l 2 , e 4 ) intersecting with square S 2 . Figure 10b shows the original triangle modified as follows: First, the intersection edge l 1 , intersection square S 2 , and the edge to which it belonged ( e 4 ) were determined. Subsequently, the projection point ϑ p t on e 4 was calculated. Finally, the vertex closest to ϑ p t in S 2 was selected, and a new triangle ( l 1 , l 3 ,   l 4 ) was used to replace triangle ( l 1 , l 2 , e 4 ).
Figure 10. Adjustment rule: (a) triangle intersecting with square and (b) triangle after modification. p t   means a b p p point. S 1 2 represent two interior squares. l 1 4 mean the edges of the triangle. ϑ p t is the projection point. The purple point is the vertex point closest to ϑ p t in S 2 .
The feature triangulation of the planar point cloud is shown in Figure 11. In Figure 11a, every b b p is connected to a certain edge of i s p , and the relation between the boundary and interior in the plane is strengthened for each P l i s . The result visually preserves the outline of the planar feature. Interior triangulation using the simplified square corner significantly decreased the number of triangles. Subsequently, we constructed a triangulated FSM in all extracted planar point clouds P l B . Our FSM in mesh form is shown in Figure 11b–d.
Figure 11. Feature triangulation: (a) triangulation of one planar point cloud, (b) all preserved vertices of triangulation, (c) the adjacency of the triangulation, and (d) the mesh model.

3.5. NSM

In robot navigation, an appropriate navigation map is necessary and favorable. Our NSM comprises an original planar point cloud P l B and a PAM. Compared with the existing NSM, the obstacles and passable area in our NSM are represented using points with coordinate dimension reduction, and the complex three-dimensional information is disregarded. Moreover, our NSM integrates P l B , a structure with a unified feature. The extracted P l B can be regarded as a feature reference map in an indoor location based on a point cloud. The constructed PAM can serve as indoor navigation and path planning. As shown in Figure 12, the flow chart of the NSM construction comprises ground extraction and map construction. The method uses R p u and P l B as inputs and generates a hybrid NSM.
Figure 12. The flow chart of navigation structure map (NSM) construction.

3.5.1. Ground Extraction

The processing pipeline of the PAM is shown in Figure 13. The input comprises P l B and R p u . In indoor scenes, typical assumptions include piecewise planar permanent structures and Manhattan world scenes in three orthogonal directions: two for the walls and one for the floors and ceilings [31]. The pipeline of the ground extraction included the following steps.
Figure 13. Pipeline of ground extraction.
(1)
Plane Grouping and Merging
P l B was grouped based on the normal direction to the plane and merged based on the distance expressed in Equation (9). Three dominant plane groups g r p i ( i = 0 ,   1 ,   2 ) were selected based on the plane quantity. Furthermore, the direction d i r i of g r p i was determined using Equation (10).
d p l = k 1 p s , n t   +   d t ( n t ) x 2 + ( n t ) y 2 +   ( n t ) z 2 + k 2 p t , n s   +   d s ( n s ) x 2 + ( n s ) y 2 +   ( n s ) z 2
d i r i = { n r l i   |   j m a x n p ( g r p j i ) ,   j = 0 , 1 , 2 n ( g r p i ) } ,
where k 1 and k 2 are two weight coefficients set to 0.5, g r p j i is the jth plane in the ith plane group, n r l is the normal to plane g r p j i , n p ( g r p j i ) is the point quantity of g r p j i , and n ( g r p i ) is the plane quantity of g r p i .
(2)
Direction Correction
In Equation (11), the plane group containing the ground, g r p g , was detected using the projection area. The correction matrix was obtained with d i r g , which is the direction of g r p g , using Equation (5). The corrected point cloud M a p C was obtained by rotating M a p V . Subsequently, the z-range ( z m i n ,   z m a x ) of M a p C was calculated.
g r p g = { g r p i   |   i m a x a r e a i , a r e a i m a x p a ( g r p j i ) ,   i = 0 , 1 , 2 ,   j = 0 , 1 , , n p ( g r p i ) } ,
where p a ( g r p j i ) is a function to compute the projection area of the jth plane in g r p i , the projection area is approximated using the area of the minimum bounding rectangle of the projection point cloud, a r e a i is the maximum projection area of the planes in g r p i , and n p ( g r p i ) is the plane quantity in the ith plane group.
(3)
Indicator Calculation
In Equation (12), the two z-ranges ( z m i n d ,   z m i n u ) and ( z m a x d ,   z m a x u ) were determined using scale parameter κ and the z range ( z m i n ,   z m a x ). Two sets of planes, s e t 1 and s e t 2 , for which the z coordinates were within ( z m i n d ,   z m i n u ) and ( z m a x d ,   z m a x u ) , respectively, were selected from g r p g . Subsequently, p l 1 c and p l 2 c , two candidate planes with maximum areas, were extracted from s e t 1 and s e t 2 , respectively. The ground and ceiling indicators were obtained using Equation (13). They comprised three states (0, 1, 2), where (a) state 0 indicates that neither the candidate ground nor ceiling is detected, (b) state 1 means that only one type is detected in the candidate ground and candidate ceiling, and (c) state 2 means that both the candidate ceiling and ground are detected. In states 0 and 1, the plane with the maximum area in g r p g is directly identified as the ground as we assume that the ceiling does not exist in this case and that the ground is the most significant plane. In state 2, the two selected candidate planes can be considered as the ground and ceiling. As both the ground and ceiling exist in the data, objects in the scene are closer to the ground than the ceiling. Furthermore, the p l k c (k = 1, 2) that is closer to the geometric center in R p u is identified as the ground. Finally, the ground plane extraction results are shown in Figure 14.
{ d =   κ ( z m a x   z m i n   )   ( z m i n d ,   z m i n u ) = ( z m i n d , z m i n + d )   ( z m a x d ,   z m a x u ) = ( z m a x d , z m a x + d )  
i n d i c a t o r = {   0 , z i p ( z m i n d ,   z m i n u ) a n d   z j p ( z m a x d ,   z m a x u ) 1 ,     z i p ( z m i n d , z m i n u ) o r     z i p ( z m a x d , z m a x u ) 2 ,     z i p ( z m i n d ,   z m i n u ) a n d     z j p ( z m a x d ,   z m a x u ) ,  
where z i p is the average z coordinate of the ith plane in g r p g , and i and j are the ith and jth planes, respectively, in g r p g .
Figure 14. Ground extraction result.

3.5.2. Map Construction

(1)
Initial Passage Area Construction
M a p V was segregated into a ground plane and space obstacles after ground extraction. They were corrected by aligning the ground with the XOY plane, and the direction-corrected ground is shown in Figure 15a. As shown in Figure 15b, Euclidean clustering was adopted to remove trivial points in the spatial obstacles. All areas were passable when no objects were above the ground. The holes in the ground were filled with contour filling in the image. As shown in Figure 15c, a binary image   I g r d based on the ground was generated based on function Λ in Equation (7). As shown in Figure 15d, the contours of the ground were extracted and filled, and then, the initial passable area in the image format was constructed.
Figure 15. Initial passable area map (PAM) and space obstacles: (a) direction-corrected ground, (b) space obstacles, (c) binary image ( I g r d ), and (d) initial PAM in image format.
(2)
PAM Construction
Because the volume of the indoor robot varied, the obstacle maps within different heights were constructed. Several obstacle maps I o b j with different heights in the image format were generated. Subsequently, I p a s s , the PAM in image format, was constructed within different heights based on Equation (14). Next, a PAM in the point cloud format was generated using Λ .
I p a s s =   I g r d I g r d     I o b j
We constructed a PAM for five height ranges (0–0.4, 0–0.8, 0–1.2, 0–1.6, and 0–2.0). A map with height range (0–1.2) is shown in Figure 16. Different areas are shown in different colors (green and red represent the passable and obstacle areas, respectively). The PAM can visually depict the position of the objects.
Figure 16. PAM with height range (0–1.2): (a) the obstacle map in binary format, (b) PAM in image format, and (c) PAM in point cloud format.
(3).
Construction of NSM
The NSM comprises the original planar point cloud described in Section 3.2 and the PAM presented Section 3.5. In robot navigation, we typically construct a global feature reference map in advance. Subsequently, the features collected in real time are exploited to match those extracted in the global feature reference map to accomplish navigation. In our study, the extracted feature is a planar structure. The PAM can be used to indicate the passable area for path planning.

4. Experiment

4.1. Platform and Data Description

As shown in Figure 17, two types of datasets were utilized in our experiment. Two simulation datasets (Figure 17a,b) were used to evaluate the performance of our proposed two-stage plane segmentation method. Four real datasets (Figure 17c–f) were used in our ISE method test. We obtained real datasets 1–3 from Park et al. [62], whereas we collected real dataset 4 using a Faro scanner. All four real datasets were collected in the connected indoor space because our method was intended for representing the indoor structure, not the space partition. Real dataset 1 was a boardroom dataset comprising a conference space and rest space. It was filled with cluttered chairs, closets, and wall structures. Real dataset 2 was an apartment dataset comprising more rooms, and real dataset 3 was a bedroom dataset with a relatively small space. Real dataset 4 was a neat exhibition room dataset with excellent planar structures. Real datasets 1–3 comprised the ground but not the ceiling; however, both structures existed in real dataset 4. A laptop with an Intel Core i7-5500U CPU @2.40 GHz/2.39 GHz and 8.0 GB of RAM was applied in our experiment.
Figure 17. Experimental datasets: (a) simulation dataset 1, (b) simulation dataset 2, (c) real dataset 1, (d) real dataset 2, (e) real dataset 3, and (f) real dataset 4.

4.2. Parameter Setting

The parameter settings for our ISE method are shown in Table 1. The parameters were classified based on to the stages of our method. (a) In preprocessing, n k 1 and r were empirically determined based on the point cloud density to process the raw point cloud. (b) In FSM construction, the parameters were set to improve the accuracy of the extracted structure and to preserve the structural details of the FSM. Figure 18 shows some of the parameters used in detail;   δ 1 and   δ 2 were approximated using Equations (15) and (16), respectively. (c) In NSM construction, the parameters used were mainly for ground extraction and PAM construction. The parameter setting in our method offers two advantages: (1) it is applicable to typical indoor dense point clouds and is independent of the scale of the scene, and (2) small changes in the parameters do not significantly affect the overall experimental result.
Table 1. Overview of parameters in proposed method.
Figure 18. Parameter setting description: (a) region growing neighborhood, (b)   δ 1 , and (c) δ 2 .
  δ 1 B C =   r k   θ 1
δ 2 2 3   r   θ 1

4.3. Experimental Results

The experimental results of four real datasets are shown in this section. The results of our proposed ISE method included an FSM and an NSM. The FSM was a mesh model generated from a plane feature. The NSM comprised a PAM and the original planar point cloud, where the PAM mainly served as obstacle indication and path planning, whereas the original planar feature point cloud was considered the localization reference. Moreover, our PAM was represented in two formats: point cloud and binary image. Some quantitative statistics of the results are listed in Table 2.
Table 2. Statistical indicators of four real datasets.

4.3.1. Real Dataset 1

Real dataset 1 was a boardroom dataset with a relatively large connection space and some clustered objects. It included some salient plane features such as a ground, conference table, wall, and cupboard. A total of 142,055 patches were constructed in the FSM (Figure 19a). The original planar point cloud (Figure 19b) contained 563 planes. The constructed PAM is shown in Figure 19c. Our NSM was composed of the original planar point cloud and the PAM. A similar display form for the NSM of real datasets 2–4 is shown in Figure 20, Figure 21 and Figure 22, respectively. The PAM between height ranges 1 and 2 differed significantly, and the change among PAMs 2–5 was insignificant. The constructed PAM conformed visually to the indoor layout.
Figure 19. Indoor structure extraction (ISE) result of real dataset 1: (a) feature structure map (FSM) result, (b) original planar point cloud, and (c) PAM.
Figure 20. ISE result of real dataset 2: (a) FSM result, (b) original planar point cloud, and (c) PAM.
Figure 21. ISE result of real dataset 3: (a) FSM result, (b) original planar point cloud, and (c) PAM.
Figure 22. ISE result of real dataset 4: (a) FSM result, (b) original planar point cloud, and (c) PAM.

4.3.2. Real Dataset 2

Real dataset 2 was an apartment dataset comprising four connected room spaces and plenty of wall structures. Compared with real dataset 1, the wall structure was more dominant. A total of 130,924 patches were constructed in the FSM (Figure 20a). The original planar point cloud (Figure 20b) contained 384 planes. Although the number of extracted planes was small, those planes had a larger area and higher integrity. As the height range increased, the bed appeared as an obstacle in the PAM, as shown in Figure 20c.

4.3.3. Real Dataset 3

Dataset 3 was a small-area bedroom dataset with one connected room space. Most of the plane features were small planar segments from the ground, wall, wardrobe, and bed. The FSM (Figure 21a) contained 94,627 patches. The original planar point cloud (Figure 21b) contained 339 planes. To accommodate the purpose of the bedroom, few obstacles were present, and the constructed PAM (Figure 21c) did not differ significantly.

4.3.4. Real Dataset 4

Dataset 4 was a reading room dataset that we collected using a Faro scanner. The merged dataset contained dense points, and the salient plane features were similar to those of real dataset 1. The most obvious difference between this dataset and real datasets 1–3 was that the former contained a large ceiling. The constructed mesh model (Figure 22a) contained 127,563 patches. The original planar point cloud (Figure 22b) contained 377 planes. As shown in Figure 22c, the two most significant obstacles in the PAM were the conference table and exhibition stand.

5. Discussion

5.1. Performance of Plane Segmentation

As show in Figure 23, both the simulation and real datasets were used to compare the plane segmentation performances between the proposed two-staged region growing and traditional region growing algorithm (TRG). Two simulation datasets (Figure 17a,b) were used for the quantitative evaluation. Simulation dataset 1 was a cube with a 0.01-m point resolution and 90,000 points on each surface. Simulation dataset 2 was a triangular prism with a 0.02-m point resolution, where each rectangle and triangle comprised 60,000 and 17,421 points, respectively. Part of real dataset 1 was used to compare the results visually.
Figure 23. Comparison of plane segmentation: (a) simulation dataset 1, (b) simulation dataset 2, and (c,d) part of real dataset 1. Red points mean unclassified points. Other color points mean different planes. In each panel, the left image is the result from traditional region growing (TRG) and the right image is the result from our method.
As shown in Figure 23, some unclassified red points appeared near the boundary or intersection of the planes in the TRG results. However, the number of unclassified points reduced visually in the results obtained using our method. Moreover, some concave and convex structures on the wall (green rectangle in Figure 23d) were not identified in the TRG but were captured successfully using our method.
The recall of the extracted plane result was estimated using Equation (17). The quantitative performance of plane segmentation for the two simulation datasets is shown in Table 3. The integrity of all the planes increased, and the average planar integrity of the two simulation datasets in our method was higher than those in the TRG, i.e., approximately 4% for the square and 6% for the triangular prism.
R e c a l l   =   P o i n t s r e c o n     P o i n t s t r u e P o i n t s t r u e ,
where P o i n t s r e c o n are the extracted plane points and P o i n t s t r u e are the real plane points.
Table 3. Performance of plane segmentation for simulation datasets.

5.2. Triangulation Performance

The constructed FSM was compared with that obtained using the traditional greedy projection triangulation (GPT) for further evaluation. In Equations (18)–(20), three statistical metrics, area mean value ( M e a n ), area standard deviation ( S t d e v ), and area range ( R a n g e ) for the patches in the mesh model, were selected to analyze the algorithm performance. These geometric attributes included the area concentration trend ( M e a n ), area variation intensity ( S t d e v ), and area variation range ( R a n g e ) of the triangular patches in the FSM. The large-volume indoor data differed from the airborne terrestrial point cloud data; it contained plenty of large-area artificial surfaces such as walls, cylinders, and related surfaces. However, a few small obstacles were present for robot localization. Therefore, the M e a n of the triangular surface area was relatively larger, and more significantly, S t d e v and R a n g e were larger. The three statistical metrics for an excellent ISE based on the point cloud should be relatively large, and they can better quantify the geometric attributes for the constructed FSM compared with the traditional data reduction ratio [25].
Moreover, the quantity of triangular patches ( Q u a n t i t y ) was used in the evaluation. A comparison of the four statistics is shown in Figure 24. As shown, the M e a n , S t d e v , and R a n g e of our method were significantly higher, whereas the Q u a n t i t y of our method was lower than that of the GPT. Our FSM attempted to capture the outline information describing the distribution of the structure and disregarded some redundant feature information. Hence, using a few large patches inside the plane and massive small patches around the boundary can not only reduce the data memory but also ensure the structure details.
M e a n = i N A r ( f i )   N
S t d e v = i N ( A r ( f i )   M e a n ) 2   N 1
R a n g e =   A r ( f i ) m a x   A r ( f i ) m i n ,
where N is the number of triangular patches in the FSM; f i is the ith triangular patch; A r ( f i ) is the area of f i ; and A r ( f i ) m a x and A r ( f i ) m i n are the maximum and minimum areas of f i respectively.
Figure 24. Statistics of patches in the mesh model: (a) M e a n and S t d e v , (b) R a n g e , and (c) Q u a n t i t y . The M e a n and S t d e v from greedy projection triangulation (GPT) were enlarged by 100 times for visual display effect.
Compared with the traditional GPT, our method processed the boundary better, improved the integrity of the planar structure, and enhanced the recall of feature segmentation. As our method can successfully capture the details of the scene structure, ISE was more robust and the appearance of the FSM was more exquisite.

6. Conclusions

Herein, we proposed a novel ISE method based on a dense unstructured point cloud that can automatically reconstruct an FSM based on a planar structure, resulting in a robust and exquisite FSM. Compared with the TRG method, our two-stage region growing method demonstrated a higher integrity of the planes. Automation of the process was improved through a logical parameter setting. The quadtree segmentation based on cluster fusion reduced redundancy. The proposed vertex assignment triangulation preserved the compact structure information around the boundary of the plane compared with GPT. The generated mesh can be applied to CAD or other software. The constructed mesh provides several functions: (1) indoor reconstruction—for some indoor complex scenes, the indoor structures should be analyzed; (2) structural repair—it strengthens the representation of the boundary feature of the structure and can serve as a basic model to preliminarily repair indoor structures; and (3) reduction in data memory—using the quadtree segmentation method, the structure of the plane point cloud can be represented with less data memory. Our ISE method can automatically reconstruct a novel form of NSM comprising the original planar point cloud and a PAM. It can identify the ground effectively in a clustered indoor scene and is not subject to the scanning coordinate, i.e., the XOY plane in the coordinate system does not need to be horizontal. In the PAM, the location of the obstacles within different heights was indicated using the points after coordinate dimension reduction. Moreover, the hybrid form of the ISE results rendered our method more applicable than the traditional single form of the ISE result.
In this study, we discovered a few challenges in the real indoor scenes: (1) our method was performed based on a dense point cloud collected using a 3D scanner or aligned via several scans. Our method may be subject to the accuracy of the original scan data; (2) the NSM result can serve as a reference feature map, and the efficiency of the construction process before real-time robot navigation can be improved; (3) movable objects in the indoor spaces must be considered; and (4) navigation may be applied to larger spaces with more rooms and corridors.
In future studies, several issues must be addressed, as follows: (1) Our ISE method only exploited the plane structure to construct the FSM; however, many types of structures exist in indoor scenes. Therefore, an FSM based on more types of structures must be investigated. (2) We should apply our NSM to real-time robot navigation using real-time perception systems in future studies. (3) A learning semantic annotation method may be investigated to label and locate the indoor structure to verify whether the constructed map results must be updated. Movable and non-movable objects in the indoor spaces must be discriminated and provided with different weights during navigation. (4) An FSS framework may be required to facilitate the use of our method for indoor robot navigation in larger spaces containing more rooms and corridors.

Author Contributions

Conceptualization, Qin Ye and Pengcheng Shi; methodology, Pengcheng Shi, Qin Ye, and Lingwen Zeng; validation, Pengcheng Shi and Lingwen Zeng; formal analysis, Qin Ye and Pengcheng Shi; investigation, Pengcheng Shi; resources, Pengcheng Shi and Qin Ye; data curation, Pengcheng Shi; writing—original draft preparation, Pengcheng Shi; writing—review and editing, Pengcheng Shi and Qin Ye; visualization, Pengcheng Shi; supervision, Qin Ye; project administration, Qin Ye; funding acquisition, Qin Ye. All authors have read and agreed to the published version of the manuscript.

Funding

The project was supported by the National Natural Science Foundation of China (grant number 41771480) and the Open Fund of Key Laboratory of Urban Land Resources Monitoring and Simulation, MNR (grant number KF-2019-04-015).

Acknowledgments

We gratefully acknowledge Jaesik Park, Qian-Yi Zhou, and Vladlen Koltun, who enabled the online availability of the experimental dataset.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Österbring, M.; Thuvander, L.; Mata, É.; Wallbaum, H. Stakeholder Specific Multi-Scale Spatial Representation of Urban Building-Stocks. ISPRS Int. J. Geo-Inf. 2018, 7, 173. [Google Scholar] [CrossRef]
  2. Lehtola, V.V.; Kaartinen, H.; Nüchter, A.; Kaijaluoto, R.; Kukko, A.; Litkey, P.; Honkavaara, E.; Rosnell, T.; Vaaja, M.T.; Virtanen, J.-P.; et al. Comparison of the Selected State-Of-The-Art 3D Indoor Scanning and Point Cloud Generation Methods. Remote Sens. 2017, 9, 796. [Google Scholar] [CrossRef]
  3. Piccinni, G.; Avitabile, G.; Coviello, G.; Talarico, C. Modeling of a Re-Configurable Indoor Positioning System Based on Software Defined Radio Architecture. In Proceedings of the 2018 New Generation of CAS (NGCAS), Valletta, Malta, 20–23 November 2018; pp. 174–177. [Google Scholar] [CrossRef]
  4. Davis, J.; Marschner, S.R.; Garr, M.; Levoy, M. Filling holes in complex surfaces using volumetric diffusion. In Proceedings of the Symposium on 3D Data Processing, Visualization, and Transmission (3DPVT), Padova, Italy, 19–21 June 2002; pp. 428–441. [Google Scholar] [CrossRef]
  5. Kim, C.; Moon, H.; Lee, W. Data management framework of drone-based 3D model reconstruction of disaster site. In Proceedings of the 23rd International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences Congress (ISPRS), Prague, Czech Republic, 12–19 July 2016; pp. 31–33. [Google Scholar] [CrossRef]
  6. Samad, S.; Lim, S.; Sara, S. Implementation of Rapid As-built Building Information Modeling Using Mobile LiDAR. In Proceedings of the 2014 Construction Research Congress (CRC): Construction in a Global Network, Atlanta, GA, USA, 19–21 May 2014; pp. 209–218. [Google Scholar] [CrossRef]
  7. Chiabrando, F.; Di Pietra, V.; Lingua, A.; Cho, Y.; Jeon, J. An Original Application of Image Recognition Based Location in Complex Indoor Environments. ISPRS Int. J. Geo-Inf. 2017, 6, 56. [Google Scholar] [CrossRef]
  8. Gupta, T.; Li, H. Indoor mapping for smart cities—An affordable approach: Using Kinect Sensor and ZED stereo camera. In Proceedings of the 2017 International Conference on Indoor Positioning and Indoor Navigation (IPIN), Sapporo, Japan, 18–21 September 2017; pp. 1–8. [Google Scholar] [CrossRef]
  9. Lan, Z.; Yew, Z.J.; Lee, G.H. Robust Point Cloud Based Reconstruction of Large-Scale Outdoor Scenes. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 9682–9690. [Google Scholar] [CrossRef]
  10. Skuratovskyi, S.; Gorovyi, I.; Vovk, V.; Sharapov, D. Outdoor Mapping Framework: From Images to 3D Model. In Proceedings of the 2019 Signal Processing Symposium (SPSympo), Krakow, Poland, 17–19 September 2019; pp. 296–399. [Google Scholar] [CrossRef]
  11. Debaditya, A.; Kourosh, K.; Stephan, W. BIM-PoseNet: Indoor camera localisation using a 3D indoor model and deep learning from synthetic images. ISPRS J. Photogramm. Remote Sens. 2019, 150, 245–258. [Google Scholar] [CrossRef]
  12. Liu, M.; Zhang, K.; Zhu, J.; Wang, J.; Guo, J.; Guo, Y. Data-Driven Indoor Scene Modeling from a Single Color Image with Iterative Object Segmentation and Model Retrieval. IEEE Trans. Vis. Comput. Graph. 2020, 26, 1702–1715. [Google Scholar] [CrossRef] [PubMed]
  13. Liu, M.; Zhang, Y.; He, J.; Guo, J.; Guo, Y. Image-Based 3D Model Retrieval for Indoor Scenes by Simulating Scene Context. In Proceedings of the 25th IEEE International Conference on Image Processing (ICIP), Athens, Greece, 7–10 October 2018; pp. 3658–3662. [Google Scholar] [CrossRef]
  14. Tran, H.; Khoshelham, K. Procedural Reconstruction of 3D Indoor Models from Lidar Data Using Reversible Jump Markov Chain Monte Carlo. Remote Sens. 2020, 12, 838. [Google Scholar] [CrossRef]
  15. Pang, M.; Luo, C.; Mei, X.; Lin, H. Acceleration of Shadowing Detection with Octree and Improved Specular Model for Indoor Propagation Using Point Cloud Data. In Proceedings of the 2018 IEEE International Conference on Computational Electromagnetics (ICCEM), Chengdu, China, 26–28 March 2018; pp. 1–3. [Google Scholar] [CrossRef]
  16. Kim, B.K. Indoor localization and point cloud generation for building interior modeling. In Proceedings of the 2013 IEEE RO-MAN, Gyeongju, Korea, 26–29 August 2013; pp. 186–191. [Google Scholar] [CrossRef]
  17. Poux, F.; Neuville, R.; Nys, G.-A.; Billen, R. 3D Point Cloud Semantic Modelling: Integrated Framework for Indoor Spaces and Furniture. Remote Sens. 2018, 10, 1412. [Google Scholar] [CrossRef]
  18. Díaz-Vilariño, L.; Martínez-Sánchez, J.; Lagüela, S.; Armesto, J.; Khoshelham, K. Door recognition in cluttered building interiors using imagery and lidar data. In Proceedings of the ISPRS Technical Commission V Symposium, Riva del Garda, Italy, 23–25 June 2014; pp. 203–209. [Google Scholar] [CrossRef]
  19. Dumitru, R.-C.; Nüchter, A. Interior reconstruction using the 3D Hough transform. In Proceedings of the 3D Virtual Reconstruction and Visualization of Complex Architectures (3D-ARCH), Trento, Italy, 25–26 February 2013; pp. 65–72. [Google Scholar]
  20. Li, J.; He, X.; Li, J. 2D LiDAR and camera fusion in 3D modeling of indoor environment. In Proceedings of the NAECON 2015—IEEE National Aerospace and Electronics Conference, Dayton, OH, USA, 15–19 June 2015; pp. 379–383. [Google Scholar] [CrossRef]
  21. Yang, L.; Cheng, H.; Su, J.; Li, X. Pixel-to-Model Distance for Robust Background Reconstruction. IEEE Trans. Circuits Syst. Video Technol. 2019, 26, 903–916. [Google Scholar] [CrossRef]
  22. Pollefeys, M.; Nister, M.; Frahm, J.M.; Akbarzadeh, A.; Mordohai, P.; Clipp, B.; Engels, C.; Gallup, D.; Kim, S.J.; Merrell, P.; et al. Detailed real-time urban 3D reconstruction from video. Int. J. Comput. Vis. 2008, 78, 143–167. [Google Scholar] [CrossRef]
  23. Wang, H.; Wang, J.; Wang, L. Online Reconstruction of Indoor Scenes from RGB-D Streams. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 3271–3279. [Google Scholar] [CrossRef]
  24. Ma, L.; Whelan, T.; Bondarev, E.; de With, P.H.N.; McDonald, J. Planar simplification and texturing of dense point cloud maps. In Proceedings of the 2013 European Conference on Mobile Robots, Barcelona, Spain, 25–27 September 2013; pp. 164–171. [Google Scholar] [CrossRef]
  25. Feichter, S.; Hlavacs, H. Planar Simplification of Indoor Point-Cloud Environments. In Proceedings of the 2018 IEEE International Conference on Artificial Intelligence and Virtual Reality (AIVR), Taichung, Taiwan, 10–12 December 2018; pp. 274–281. [Google Scholar] [CrossRef]
  26. Adan, A.; Huber, D. 3D Reconstruction of Interior Wall Surfaces under Occlusion and Clutter. In Proceedings of the 2011 International Conference on 3D Imaging, Modeling, Processing, Visualization and Transmission (3DIMPVT), Hangzhou, China, 16–19 May 2011; pp. 275–281. [Google Scholar] [CrossRef]
  27. Oesau, S.; Lafarge, F.; Alliez, P. Indoor Scene Reconstruction using Feature Sensitive Primitive Extraction and Graph-cut. ISPRS J. Photogramm. Remote Sens. 2014, 90, 68–82. [Google Scholar] [CrossRef]
  28. Shui, W.; Liu, J.; Ren, P.; Maddock, S.; Zhou, M. Automatic planar shape segmentation from indoor point clouds. In Proceedings of the 15th ACM SIGGRAPH Conference on Virtual-Reality Continuum and Its Applications in Industry, VRCAI, Zhuhai, China, 3–4 December 2016; pp. 363–372. [Google Scholar] [CrossRef]
  29. Georgios-Tsampikos, M.; Renato, P. Bayesian graph-cut optimization for wall surfaces reconstruction in indoor environments. Vis. Comput. 2017, 33, 1347–1355. [Google Scholar] [CrossRef]
  30. Zhou, K.; Gorte, B.; Zlatanova, S. Exploring Regularities for improving façade reconstruction from point clouds. In Proceedings of the 23rd International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences Congress (ISPRS), Prague, Czech Republic, 12–19 July 2016; pp. 749–755. [Google Scholar] [CrossRef]
  31. Ma, L.; Favier, R.; Do, L.; Bondarev, E.; de With, P.H.N. Plane segmentation and decimation of point clouds for 3D environment reconstruction. In Proceedings of the 2013 IEEE 10th Consumer Communications and Networking Conference (CCNC), Las Vegas, NV, USA, 11–14 January 2013; pp. 43–49. [Google Scholar] [CrossRef]
  32. Turner, E.; Zakhor, A. Watertight Planar Surface Meshing of Indoor Point-Clouds with Voxel Carving. In Proceedings of the 2013 International Conference on 3D Vision, Seattle, WA, USA, 29 June–1 July 2013; pp. 41–48. [Google Scholar] [CrossRef]
  33. Dimitrov, A.; Golparvar-Fard, M. Segmentation of building point cloud models including detailed architectural/structural features and MEP systems. Autom. Constr. 2015, 51, 32–45. [Google Scholar] [CrossRef]
  34. Poux, F.; Billen, R. Voxel-based 3D point cloud semantic segmentation: Unsupervised geometric and relationship featuring vs deep learning methods. ISPRS Int. J. Geo-Inf. 2019, 8, 213. [Google Scholar] [CrossRef]
  35. Papon, J.; Abramov, A.; Schoeler, M.; Worgotter, F. Voxel cloud connectivity segmentation—Supervoxels for point clouds. In Proceedings of the 2013 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Portland, OR, USA, 23–28 June 2013; pp. 2027–2034. [Google Scholar] [CrossRef]
  36. Vo, A.V.; Truong-Hong, L.; Laefer, D.F.; Bertolotto, M. Octree-based region growing for point cloud segmentation. ISPRS J. Photogramm. Remote Sens. 2015, 104, 88–100. [Google Scholar] [CrossRef]
  37. Armeni, I.; Sener, O.; Zamir, A.R.; Jiang, H.; Brilakis, I.; Fischer, M.; Savarese, S. 3D Semantic Parsing of Large-Scale Indoor Spaces. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 1534–1543. [Google Scholar] [CrossRef]
  38. Ochmann, S.; Vock, R.; Wessel, R.; Tamke, M.; Klein, R. Automatic generation of structural building descriptions from 3D point cloud scans. In Proceedings of the 9th International Conference on Computer Graphics Theory and Applications, Lisbon, Portugal, 5–8 January 2014; pp. 120–127. Available online: https://cg.cs.uni-bonn.de/aigaion2root/attachments/GRAPP_2014_54_CR.pdf (accessed on 17 June 2020).
  39. Tchuinkou Kwadjo, D.; Tchinda, N.; Bobda, C.; Menadjou, N.; Fotsing, C.; Nziengam, N. From PC2BIM: Automatic Model generation from Indoor Point Cloud. In Proceedings of the 13th International Conference on Distributed Smart Cameras, Trento, Italy, 9–11 September 2019; pp. 1–6. [Google Scholar] [CrossRef]
  40. Yang, F.; Li, L.; Su, F.; Li, D.; Zhu, H.; Ying, S.; Zuo, X.; Tang, L. Semantic decomposition and recognition of indoor spaces with structural constraints for 3D indoor modelling. Autom. Constr. 2019, 106, 319–339. [Google Scholar] [CrossRef]
  41. Shi, W.; Ahmed, W.; Li, N.; Fan, W.; Xiang, H.; Wang, M. Semantic Geometric Modelling of Unstructured Indoor Point Cloud. ISPRS Int. J. Geo-Inf. 2019, 8, 9. [Google Scholar] [CrossRef]
  42. Wang, C.; Hou, S.; Wen, C.; Gong, Z.; Li, Q.; Sun, X.; Li, J. Semantic line framework-based indoor building modeling using backpacked laser scanning point cloud. ISPRS J. Photogramm. Remote Sens. 2018, 143, 150–166. [Google Scholar] [CrossRef]
  43. Xiong, X.; Huber, D. Using Context to Create Semantic 3D Models of Indoor Environments. In Proceedings of the 2010 21st British Machine Vision Conference, Aberystwyth, Wales, UK, 31 August–3 September 2010; pp. 45.1–45.11. [Google Scholar] [CrossRef]
  44. Cui, Y.; Li, Q.; Yang, B.; Xiao, W.; Chen, C.; Dong, Z. Automatic 3-D Reconstruction of Indoor Environment With Mobile Laser Scanning Point Clouds. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 3117–3130. [Google Scholar] [CrossRef]
  45. Xiong, X.; Adan, A.; Akinci, B.; Huber, D. Automatic Creation of Semantically Rich 3D Building Models from Laser Scanner Data. Autom. Constr. 2013, 31, 325–337. [Google Scholar] [CrossRef]
  46. Yang, L.; Worboys, M. Generation of navigation graphs for indoor space. Int. J. Geogr. Inf. Sci. 2015, 29, 1737–1756. [Google Scholar] [CrossRef]
  47. Brawn, G.; Nagel, C.; Zlatanova, S.; Kolbe, T. Modelling 3D Topographic Space against Indoor Navigation Requirements. In Proceedings of the 7th International Workshop on 3D Geoinformation, Quebec City, QC, Canada, 16–17 May 2012; pp. 1–22. [Google Scholar] [CrossRef]
  48. Nikoohemat, S.; Diakité, A.; Zlatanova, S.; Vosselman, G. Indoor 3D Modeling and Flexible Space Subdivision from Point Clouds. In Proceedings of the 4th ISPRS Geospatial Week, Enschede, The Netherlands, 10–14 June 2019; pp. 285–292. [Google Scholar] [CrossRef]
  49. Diakité, A.A.; Zlatanova, S. Spatial subdivision of complex indoor environments for 3D indoor navigation. Int. J. Geogr. Inf. Sci. 2018, 32, 213–235. [Google Scholar] [CrossRef]
  50. Zlatanova, S.; Liu, L.; Sithole, G. A conceptual framework of space subdivision for indoor navigation. In Proceedings of the 5th ACM SIGSPATIAL International Workshop on Indoor SpatialAwareness, Orlando, FL, USA, 5 November 2013; pp. 37–41. [Google Scholar] [CrossRef]
  51. Taneja, S.; Akinci, B.; Garrett, J.; Soibelman, L. Algorithms for automated generation of navigation models from building information models to support indoor map-matching. Autom. Constr. 2016, 61, 24–41. [Google Scholar] [CrossRef]
  52. Flikweert, P.; Peters, R.; Díaz-Vilariño, L.; Voûte, R.; Staats, B. Automatic Extraction of A Navigation Graph Intended for IndoorGML From An Indoor Point Cloud. In Proceedings of the 4th ISPRS Geospatial Week, Enschede, The Netherlands, 10–14 June 2019; pp. 271–278. [Google Scholar] [CrossRef]
  53. Boguslawski, P.; Mahdjoubi, L.; Zverovich, V.; Fadli, F. Automated construction of variable density navigable networks in a 3D indoor environment for emergency response. Autom. Constr. 2016, 72, 115–128. [Google Scholar] [CrossRef]
  54. Li, F.; Wang, H.; Akwensi, P.H.; Kang, Z. Construction of Obstacle Element Map Based on Indoor Scene Recognition. In Proceedings of the 4th ISPRS Geospatial Week, Enschede, The Netherland, 10–14 June 2019; pp. 819–825. [Google Scholar] [CrossRef]
  55. Nakagawa, M.; Nozaki, R. Geometrical Network Model Generation Using Point Cloud Data for Indoor Navigation. In Proceedings of the 2018 ISPRS TC IV Mid-Term Symposium on 3D Spatial Information Science—The Engine of Change, Delft, The Netherlands, 1–5 October 2018; pp. 141–146. [Google Scholar] [CrossRef]
  56. Liu, L.; Zlatanova, S. Generating Navigation Models from Existing Building Data. In Proceedings of the 2013 ISPRS Acquisition and Modelling of Indoor and Enclosed Environments, Cape Town, South Africa, 11–13 December 2013; pp. 19–25. [Google Scholar] [CrossRef]
  57. Abdullah, A.; Sisi, Z.; Peter, V.; Ki-Joune, L. Improved and More Complete Conceptual Model for the Revision of IndoorGML. In Proceedings of the 10th International Conference on Geographic Information Science (GIScience 2018), Melbourne, Australia, 28–31 August 2018. [Google Scholar] [CrossRef]
  58. Walton, L.A.; Worboys, M. A Qualitative Bigraph Model for Indoor Space. In Proceedings of the 7th International Conference on Geographic Information Science (GIScience 2012), Columbus, OH, USA, 18–21 September 2012; pp. 226–240. [Google Scholar] [CrossRef]
  59. Schnabel, R.; Wahl, R.; Klein, R. Efficient RANSAC for point-cloud shape detection. Comput. Graph. Forum 2007, 26, 214–226. [Google Scholar] [CrossRef]
  60. Okorn, B.E.; Xiong, X.; Akinci, B.; Huber, D. Toward automated modeling of floor plans. In Proceedings of the Symposium on 3D Data Processing, Visualization and Transmission, Paris, France, 17–20 May 2010; Available online: https://ri.cmu.edu/pub_files/2010/5/2009%203DPVT%20plan%20view%20modeling%20v13%20(resubmitted).pdf (accessed on 23 June 2020).
  61. Li, M.; Wonka, P.; Nan, L. Manhattan-world urban reconstruction from point clouds. In Proceedings of the European Conference on Computer Vision (ECCV), Amsterdam, The Netherlands, 11–14 October 2016; pp. 54–69. [Google Scholar] [CrossRef]
  62. Park, J.; Zhou, Q.; Koltun, V. Colored point cloud registration revisited. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 143–152. [Google Scholar] [CrossRef]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.