Skip to Content
IJGIISPRS International Journal of Geo-Information
  • Article
  • Open Access

20 February 2026

Airborne LiDAR Point Cloud Building Reconstruction Based on Planar Optimal Combination and Feature Line Constraints

,
,
,
,
and
1
School of Civil Engineering and Geomatics, Shandong University of Technology, Zibo 255000, China
2
Hubei Luojia Laboratory, Wuhan 430079, China
*
Author to whom correspondence should be addressed.

Abstract

This paper proposes a building reconstruction framework for airborne LiDAR data to address the challenge of automated modeling under conditions of uneven point cloud density and missing vertical walls, generating high-precision and structurally compact 3D building models. The method first combines adaptive resolution hypervoxels with a global graph cut optimization strategy to extract precise roof plane primitives from sparse point clouds of buildings. Subsequently, it infers building facades and internal vertical walls based on point cloud projection contours and height change detection, thereby completing the wall structures commonly missing in airborne LiDAR data. Finally, a feature line constraint term is introduced into the hypothesis-and-selection-based reconstruction framework to guide the structural optimization of candidate planes, ensuring the reconstructed model closely matches the actual building geometry. The proposed method was evaluated on multiple public airborne LiDAR datasets, demonstrating its effectiveness through qualitative and quantitative comparisons with various state-of-the-art approaches.

1. Introduction

The digitization of urban scenes represents a significant research direction in fields such as computer vision, remote sensing, and computer graphics [1,2]. As a critical infrastructure for digital twin cities, 3D building models have been widely applied in numerous domains, including urban planning [3], navigation and positioning [4,5], disaster early warning [6], energy assessment [7], and environmental analysis [8].
Among various data sources, airborne LiDAR scanning offers distinct advantages: it efficiently captures three-dimensional structural information of building roofs over large areas, while its emitted pulses can penetrate vegetation or shadow-obscured areas, thereby supplementing building information that is difficult to detect and reconstruct using optical imagery [9]. Consequently, urban 3D building reconstruction based on airborne LiDAR point clouds has garnered extensive research attention [10,11,12,13,14]. However, existing technical methods still face significant challenges in achieving automated reconstruction of highly accurate and structurally compact building models. First, the overall distribution of airborne LiDAR point clouds is sparse and unevenly dense, complicating the precise extraction of roof plane primitives. Second, due to sensor field-of-view and flight path limitations, critical structural information, such as vertical building walls, is often incomplete or even absent in airborne LiDAR point clouds. Furthermore, the diverse forms and complex structures of urban buildings further increase the difficulty of model topological reconstruction.
To address these issues, this paper proposes a building reconstruction method based on planar optimization combinations and feature line constraints for automatically generating compact and effective 3D building models from airborne LiDAR point clouds. The method first converts the discrete point cloud into a voxel, then combines a global graph cut strategy to extract and optimize the roof plane primitives. Subsequently, projected contours are extracted from the building point cloud, and height change point pairs are detected to infer the building’s exterior walls and internal vertical walls. Finally, the method incorporates roof feature lines as structural prior into the hypothesis-and-selection-based reconstruction framework [15] to guide the screening and optimization of candidate surface sets, thereby improving topological accuracy and enhancing detail recovery. The main contributions of this paper are as follows:
  • To address the issues of uneven density and unstable boundaries in airborne LiDAR point clouds, this study proposes a roof plane segmentation method combining adaptive resolution supervoxels with graph cut re-labeling. This approach enhances adaptability to density variations while effectively suppressing jagged boundaries and oversegmentation problems.
  • Introduces a vertical plane inference strategy based on height change detection to effectively restore missing vertical wall information prevalent in airborne LiDAR data.
  • Introduces a feature line constraint term within the hypothesis-and-selection-based reconstruction framework, encouraging the prioritization of planar combinations connected to roof feature lines to achieve reconstruction models that better align with actual building geometries.

3. Methodology

3.1. Overview

This method takes point clouds of individual buildings captured by airborne LiDAR as input and outputs a three-dimensional polygonal building surface model that satisfies watertight and 2-manifold properties. The current implementation exports results in the standard OBJ polygon mesh format. The overall workflow is illustrated in Figure 1.
Figure 1. Overall Methodology Flowchart. (a) Input point cloud; (b) Roof segmentation plane; (c) Roof feature lines; (d) Inference of vertical walls; (e) Hypothesis of candidate surface set; (f) Reconstructed model.
The method primarily consists of three stages: First, a planar segmentation strategy based on adaptive resolution voxels and globally optimized graph cuts is employed to extract roof plane primitives. Subsequently, building facades and internal vertical walls are inferred using projected contours and height change information from the point cloud. Finally, under the hypothesis-and-selection-based reconstruction framework, a hypothesis set of building surfaces is generated based on the extracted roof planes and inferred vertical planes. The optimal subset is then selected through optimization to produce the final model. In this process, a feature line constraint term is further introduced. Roof feature lines are generated by intersecting adjacent roof planes, and a matching relationship is established between candidate plane intersections and the feature lines, thereby guiding the optimization process to prioritize candidate plane intersections that match the feature lines and their corresponding plane combinations.

3.2. Roof Plane Segmentation

Due to the uneven density distribution of airborne LiDAR point clouds, traditional planar detection methods such as RANSAC and region growing are prone to oversegmentation and undersegmentation and struggle to preserve planar boundaries effectively. To extract complete and accurate planar primitives from building roof point clouds for subsequent reconstruction, this paper proposes a roof plane segmentation method combining adaptive resolution hypervoxels with global graph cut optimization: First, a structurally robust coarse segmentation is obtained via adaptive resolution hypervoxels. Then, hierarchical clustering enhances planar completeness. Finally, global graph cuts are applied to relabel point-level planar tags, correcting jagged boundaries and local oversegmentation. The overall workflow is illustrated in Figure 2.
Figure 2. Roof Planar Segmentation. (a) Input building point cloud; (b) Hypervoxel segmentation result; (c) Clustering and merging result; (d) Graph cut optimization result.

3.2.1. Adaptive Resolution Hypervoxel Coarse Segmentation

Given the disordered nature of point cloud data, this paper introduces a representative point-driven hypervoxel coarse segmentation method [47]. By incorporating local distance constraints and an adaptive balancing mechanism for representative point counts into the energy function, this approach enables hypervoxel resolution to automatically adjust based on local point density. This achieves boundary-friendly and structurally robust hypervoxel coarse segmentation, thereby reducing the risk of cross-plane undersegmentation. The energy function is defined as follows:
min E ( Z ) = i = 1 N j = 1 N z i j D p i , p j + λ   |   C ( Z ) K   | ,       s . t .   z i j = { 0 , 1 } ,   i = 1 N z i j = 1 , j
where z i j indicates whether the neighboring point p j is represented by the representative point p i ; D p i , p j represents the distance between the representative point and its neighboring points; λ is the parameter balancing distance and voxel count; C ( Z ) represents the number of representative points; K denotes the preset voxel count.

3.2.2. Hierarchical Clustering Merge

As shown in Figure 2b, a roof surface is divided into multiple super-voxels of varying scales. Merging these super-voxels yields the complete roof surface. First, an adjacency graph G = ( V , E ) is constructed among the super-voxels, where each node v V represents a super-voxel and each edge e E connects two adjacent super-voxels. Subsequently, hierarchical clustering is performed on this graph to iteratively merge hypervoxels. In each iteration, the edge e ( p , q ) with the minimum mean squared error (MSE) is selected, merging its corresponding two hypervoxels into a new node while updating the graph structure. This process is repeated until the minimum MSE exceeds a preset threshold T m , at which point the iterative aggregation process terminates. By progressively enhancing the integrity of roof planes through hierarchical clustering merging, more reliable initial planes are provided for subsequent global graph cut optimization.

3.2.3. Global Graph Cutting Optimization

As indicated by the red rectangle in Figure 2c, after voxel coarse segmentation and cluster merging, some roof surfaces still exhibit jagged boundaries and over-segmentation issues. This paper addresses these problems by performing point-level global relabeling on the clustered plane labels through modeling and solving a global energy minimization problem. The global energy function is defined as:
E ( L ) = p P D p L p + p , q N δ p , q L p L q
where data item D p L p represents the geometric distance from point p to plane L p , ensuring that each point tends to be assigned to the plane with the shortest distance:
D p L p = a L P x p + b L P y p + c L P z p + d L P a L P 2 + b L P 2 + c L P 2
The plane equation corresponding to L p is a L P x p + b L P y p + c L P z p + d L P = 0 .
The smoothing term δ p , q L p , L q penalizes cases where adjacent points are assigned to different planes, thereby enforcing spatial continuity and eliminating jagged boundaries.
δ p , q L p , L q = 0     L p = L q exp ( x p x q 2 + y p y q 2 + z p z q 2 )   L p L q
After constructing the optimization function, the α β exchange technique based on multi-label cut minimizes the objective function E ( L ) , thereby obtaining an approximate global optimum solution. Figure 2d illustrates the optimization results, demonstrating that jagged boundaries and over-segmentation issues have been effectively resolved.
Through the aforementioned coarse-to-fine planar clustering and optimization workflow, the proposed method effectively mitigates issues such as oversegmentation, undersegmentation, and boundary jaggedness. This enhances the accuracy and completeness of aerial LiDAR point cloud planar segmentation, thereby providing a more reliable topological planar foundation for subsequent building reconstruction.

3.3. Vertical Wall Inference

In airborne LiDAR point clouds, due to the limited scanning field of view of the sensors, vertical building walls are often sparsely sampled or even completely absent. However, vertical walls play a crucial role in representing building structures, encompassing not only the main exterior walls connecting the roof to the ground but also internal vertical walls located between multi-level roofs. To address this, this paper combines projected contours and height change information from point clouds to jointly infer both internal and external vertical walls of buildings. The overall workflow is illustrated in Figure 3.
Figure 3. Vertical Plane Inference. (a) Input point cloud; (b) Alpha-shape outer boundary; (c) Regularized outer contour; (d) 2D footprint; (e) Height change points; (f) Inner wall segments; (g) Inner and outer wall segments; (h) Inner and outer vertical planes.

3.3.1. Vertical Exterior Wall Inference

First, the point cloud of the single building is projected onto a two-dimensional horizontal plane, and its initial outer boundary is extracted using the α-shape method, as shown in Figure 3b. Subsequently, segmented least-squares line fitting is performed on the boundary point sequence, and collinear points within a distance threshold are merged using a region-growing strategy to obtain an initial set of line segments. Building upon this, directional clustering is performed on the extracted line segments, and the orientations within each cluster are adjusted to the average direction. After completing the orientation adjustment, approximately collinear and approximately orthogonal line segments are further refined to be exactly collinear and orthogonal, respectively, ultimately yielding a regularized building outer footprint, as illustrated in Figure 3c.
Based on this regularized contour, the vertical exterior walls of the building can be obtained by extrapolating along the vertical direction. Additionally, the contour polygon represents the building’s two-dimensional footprint, as illustrated in Figure 3d. Therefore, this paper enables direct model reconstruction without requiring the building footprint as prior information.

3.3.2. Vertical Interior Wall Inference

In airborne LiDAR point clouds, interior vertical walls of buildings typically manifest as significant height changes. Based on this characteristic, height change point pairs are first detected from the original point clouds, and the spatial distribution of interior vertical walls is subsequently inferred. Let the input building point cloud be P = p i = x i , y i , z i . For any two points p i , p j P , their two-dimensional horizontal distance and height difference are defined as:
d i j = x i x j 2 + y i y j 2 h i j = z i z j
when the following conditions are satisfied, point p i , p j is identified as a height change point pair:
d i j < ε d , h i j > ε h
where ε d denotes the horizontal neighborhood search radius, and ε h represents the height change threshold.
By performing a neighborhood search on the building point cloud under the aforementioned conditions, a set of height change point pairs H can be obtained. Figure 3e illustrates the detection results: red points indicate upper-layer height change points, while blue points denote corresponding lower-layer height change points. Subsequently, the upper-layer height change points are projected onto a two-dimensional plane. Piecewise least-squares line fitting is then applied to extract two-dimensional line segments representing the positions of internal walls, as shown in Figure 3f. Considering that interior walls in building layouts are typically parallel or perpendicular to the outer footprint, an orientation regularization constraint is further imposed on interior wall segments to enforce parallelism or orthogonality with the principal directions of the projected outer footprint, as shown in Figure 3g.
Finally, by extrapolating the regularized outer contour and internal wall segments along the perpendicular direction, the building’s inner and outer vertical planes are obtained, as shown in Figure 3h. This method effectively supplements missing vertical wall information in airborne LiDAR point clouds, providing reliable planar primitives for subsequent closed reconstruction and topological optimization of building 3D models.

3.4. Building Model Reconstruction

This paper employs the hypothesis-and-selection-based framework for building model reconstruction: first, roof segmentation planes and internal vertical planes are extended in space and intersected pairwise to generate a set of hypothetical candidate surfaces (see Figure 1e); subsequently, an optimal subset is selected from this set, and the result is clipped by external vertical planes to obtain a closed building surface model. To further encourage the generation of physically plausible building models, this paper introduces a novel energy term. The optimization objective for the candidate surface patch set F is defined as:
F = arg min X   λ d E d + λ c E c + λ l E l
where X = x i x i { 0 , 1 } is the binary selection variable for candidate surfaces. E d is the data fitting term, which encourages the selection of surface patches supported by more points; E c is the model complexity term, used to suppress overly complex surface combinations (see [15] for details). E l is the feature line constraint term proposed in this paper, designed to enhance the topological rationality of roof reconstruction at the structural level.
In complex architectural scenarios, the hypothesis-and-selection framework generates a vast number of candidate surfaces. Relying solely on the optimization process of and can easily lead to topological errors, resulting in surface combinations that “appear plausible but fail to match the actual roof geometry.” To address this, this paper establishes a matching relationship between candidate plane intersections and roof feature lines, leveraging the structural prior of roof feature lines to guide the screening and optimization of candidate planes.

3.4.1. Matching Relationship Construction

Let the set of candidate plane intersections be denoted as e = e 1 , e 2 , , e m , and the set of roof feature lines as l = l 1 , l 2 , , l n . As shown in Figure 4a, the roof feature lines are obtained by intersecting adjacent roof segmentation planes and are trimmed via point cloud projection to better approximate their actual lengths. This approach can partially reflect key geometric structures such as ridge lines and valley lines. The candidate plane intersection lines participating in matching are generated by extending the roof segmentation planes and intersecting them in pairs. These lines are then clipped by a vertical plane into segments of varying lengths, as shown in Figure 4b. Theoretically, the roof feature line should be a subset of the candidate plane intersection lines, being completely collinear with some candidates in space but differing only in length. Figure 4c shows the overlay effect of the two, where the green line represents the roof feature line. Based on this, if the candidate intersection line e i and feature line l k simultaneously satisfy the following joint conditions, e i is considered to match l k :
e i l k e i l k e i l k e i τ o
where e i l k denotes that the two lines are collinear (i.e., share the same direction); e i l k e i represents the proportion of the candidate intersection line covered by the feature line; τ o denotes the overlap rate threshold, set at 0.9 in this study’s experiments to ensure only high-confidence matches are retained.
Figure 4. Schematic diagram of roof feature lines and candidate plane intersections. (a) Roof feature lines; (b) Candidate plane intersections; (c) Overlay of both (green indicates roof feature lines).

3.4.2. Feature Line Constraint Item

E l is defined based on the aforementioned matching relationship and is used to constrain the selection process of candidate planes at the structural level:
E l = 1 | E | i = 1 | E | i n t e r s e c t e i
where | E | denotes the number of all pairwise intersecting surfaces within the candidate surface set; i n t e r s e c t e i is a binary indicator function denoting that an edge connects two intersecting surfaces e i : if e i matches a specific roof feature line l k , this indicator takes 0; otherwise, it takes 1. This term is designed to increase the likelihood of topologically correct roof plane selection by biasing the cost function toward plane combinations connected to roof feature lines, thereby reducing structural misselections caused by purely data-driven fitting and complexity control. It should be noted that this feature line constraint, as a soft constraint, does not enforce the selection of planes. Instead, it imposes a structural bias on the optimization process by reducing the corresponding energy cost. The final outcome remains jointly determined by the data fitting term and the model complexity term.
Figure 5 illustrates the difference in local reconstruction when the feature line constraint term E l is introduced. The roof appendage within the red rectangle is prone to being discarded when the feature line constraint is not applied, due to its limited support points on the segmentation plane and significantly increased model complexity. However, introducing E l preserves this appendage.
Figure 5. Effect of Feature Line Constraint E l on Reconstruction Results. (a) The input consists of segmented roof planes and the feature lines generated by their intersections; (b) Reconstruction result without E l ; (c) Reconstruction result with E l incorporated.
In summary, this paper introduces structural priors based on roof feature lines within the hypothesis-selection framework. This approach enables the reconstruction process to balance point cloud fitting with model simplicity while closely following critical structural edges such as ridges and valleys. Consequently, it yields architectural surface models that more accurately reflect the true geometric and topological relationships of the roof.

4. Experiments

4.1. Experimental Setup

The methodology of this study was implemented in C++ (C++17), with each module’s functionality developed using CGAL 5.4.1 and PCL 1.14.0 libraries. Experiments were conducted on a device equipped with an Intel Core i9-12900H processor at 2.50 GHz and 16 GB of memory.

4.1.1. Test Dataset

We tested the proposed method on three publicly available airborne LiDAR point cloud datasets for qualitative and quantitative evaluation.
  • AHN3 [48]: AHN3 is a large-scale airborne LiDAR point cloud dataset covering the entire Netherlands, collected by the Dutch National Topographic Service with an average point density of 8 points/m2. The dataset provides multi-class semantic labels, including buildings, trees, ground, and water bodies, enabling direct extraction of building point clouds.
  • DALES [49]: DALES is a large-scale aerial point cloud dataset collected in Surrey, Canada. Data were acquired using a Riegl Q1560 dual-channel system at an altitude of approximately 1300 m, a speed of 72 m/s, and an average point density of 50 points/m2. All buildings are provided with instance-level annotations.
  • Vaihingen [50]: Vaihingen was released by the German Society for Photogrammetry and Remote Sensing (ISPRS) and collected in Vaihingen, Germany. The test area in this paper belongs to the benchmark data of the ISPRS “Urban Classification and 3D Building Reconstruction Test Project.” Data was acquired using a Leica ALS50 system with a 45° field of view, at an altitude of approximately 500 m, and an average point density of 4 points/m2. Building point clouds can be directly obtained through the provided classification labels.

4.1.2. Evaluation Indicators

To quantitatively evaluate the proposed method in this paper, we examine it from two aspects: plane segmentation and model reconstruction. According to [51], the evaluation metrics for plane segmentation include: Completeness ( C m ), Correctness ( C r ), and Quality ( Q l ), defined as follows:
C m = | T P | | T P | + | F N | C r = | T P | | T P | + | F P | Q l = | T P | | T P | + | F P | + | F N |
where T P (True Positive) refers to correctly segmented roof surfaces; F N (False Negative) refers to missed roof surfaces; F P (False Positive) refers to incorrectly segmented roof surfaces.
During the model reconstruction process, since no ground-truth reconstruction results are available for the buildings in the three datasets, the commonly used accuracy metric, Root Mean Square Error (RMSE), is employed to evaluate the quality of each reconstructed model. In the context of surface reconstruction, given a point cloud set P = p i i = 1 N , RMSE is defined as the square root of the mean squared Euclidean distance from the points to the reconstructed model:
RMSE = 1 N i = 1 N p i p ^ i 2
where p i is the i -th point in the point cloud; p ^ i is the projection point of point p i onto the surface of the reconstruction model; and p i p ^ i is the Euclidean distance.

4.1.3. Parameter Settings

The method described in this paper involves several parameters whose values are determined based on the point cloud density of the dataset and experimental experience. During the planar segmentation stage, the mean squared error threshold T m for hypervoxel clustering and merging is set to 0.25. In the vertical plane inference stage, the height difference threshold ε h for height change detection is set to 0.5 m. The horizontal neighborhood radius ε d is adaptively adjusted based on the point cloud density of different datasets, set to twice the average inter-point distance. During the building model reconstruction stage, the feature line matching overlap rate τ o is set to 0.9, and the weight λ l of the feature line constraint term is set to 0.08.

4.2. Planar Segmentation Results

To validate the effectiveness and robustness of the proposed planar segmentation algorithm, comparative experiments were conducted against three classical plane detection methods: the RANSAC algorithm [19], the Region Growing (RG) algorithm [21], and the Global Optimization (GO) algorithm [52]. Partial segmentation results of the four methods on different datasets are shown in Figure 6. The figure displays six representative building instances: the top two rows originate from the AHN3 dataset, the middle two rows from the DALES dataset, and the bottom two rows from the Vaihingen dataset. From left to right: (a) Input building point cloud, (b) Manually annotated reference segmentation result, (c) Segmentation result from the proposed method, (d) Segmentation result from the RANSAC method, (e) Segmentation result from the RG method, (f) Segmentation result from the GO method.
Figure 6. Roof plane segmentation results. (a) Input point cloud; (b) Manually annotated reference segmentation result; (c) Segmentation result from the proposed method; (d) Segmentation result from the RANSAC method; (e) Segmentation result from the RG method; (f) Segmentation result from the GO method.
Overall, the proposed method demonstrates significant advantages in segmentation completeness, boundary accuracy, and adaptability to complex roof structures. For instance, as observed in Figure 6d, while the RANSAC method generates relatively clear planar boundaries, it tends to produce oversegmentation in sparse point cloud regions. In Figure 6e, the RG method struggles to form regular and clear planar boundaries at different plane intersections. Also, it tends to exhibit oversegmentation or undersegmentation in complex roof structure areas. In Figure 6f, although the GO method achieves high overall segmentation accuracy with relatively smooth boundaries, it is prone to undersegmentation in regions with indistinct boundary features; simultaneously, oversegmentation still occurs in low-density point cloud areas.
In contrast, the proposed method maintains both segmentation accuracy and planar integrity under uneven point cloud density. This advantage stems from converting discrete point clouds into structurally stable hypervoxels, effectively mitigating risks of oversegmentation and undersegmentation. Concurrently, incorporating hierarchical clustering and global graph cut optimization progressively enhances planar integrity and boundary precision.
Table 1 presents quantitative evaluation results from three datasets, with 10 buildings selected from each dataset for statistical analysis. A comprehensive review of all evaluation metrics reveals that the proposed method outperforms other comparison algorithms across different datasets in terms of segmentation completeness, accuracy, and overall quality. This further validates the effectiveness and generalization capability of the proposed planar segmentation method in diverse architectural scenarios.
Table 1. Quantitative evaluation results of different plane segmentation methods. C m , C r and Q l represent Completeness, Correctness and Quality defined in Equation (10).
To further analyze the impact of planar segmentation results on subsequent building reconstruction, the segmentation outputs from different algorithms were input into the reconstruction framework proposed in this paper to obtain corresponding reconstruction results, as shown in Figure 7.
Figure 7. Model reconstruction results using different planar segmentation methods. (a) Segmentation and reconstruction results of the proposed method; (b) Segmentation and reconstruction results of the RANSAC method; (c) Segmentation and reconstruction results of the RG method; (d) Segmentation and reconstruction results of the GO method.
As shown in Figure 7a, the reconstructed roof structure generated from the planar segmentation results based on the proposed method exhibits regular and reasonable features, indicating that this segmentation approach can provide accurate and reliable topological planar primitives for subsequent reconstruction processes. Within the red-boxed area in Figure 7b, RANSAC incorrectly classifies distinct roof surfaces with significant height differences as belonging to the same plane, resulting in topological errors in the reconstructed model. In Figure 7c, the RG method struggles to generate clear plane boundaries, resulting in numerous small, fragmented surfaces at the intersections of multiple roof planes in the reconstruction. Within the red-boxed area of Figure 7d, the GO method exhibits a small number of oversegmented and undersegmented plane primitives, causing partial roof structures to be reconstructed incorrectly.
In summary, the proposed planar segmentation algorithm effectively enhances segmentation accuracy while preserving planar integrity, providing stable and reliable topological planar primitives for subsequent model reconstruction.

4.3. Model Reconstruction Results

Figure 8a shows a large test area within the AHN3 dataset, where red points represent building point clouds. After manually cropping the building point clouds in this area into individual buildings, they were input into the reconstruction process proposed in this paper, yielding the corresponding reconstruction results shown in Figure 8b. It can be observed that for most buildings, the proposed method generates structurally complete and compact 3D building models. Within this test area, a total of 273 buildings were included, with an overall reconstruction time of approximately 17 min. The average reconstruction time per building was about 3.77 s. These results demonstrate that the method proposed in this paper exhibits good processing efficiency and scalability in large-scale scenarios.
Figure 8. Large-scale building model reconstruction results from the AHN3 dataset. (a) Input point cloud; (b) Building model.
To better illustrate model details, representative buildings from the three test datasets are selected and magnified in Figure 9. Despite issues such as uneven point cloud density and data gaps in the input, as well as significant variations in roof structures across different buildings, the proposed method successfully reconstructs visually complete and plausible 3D building models. This demonstrates the effectiveness of the method in extracting roof planes and inferring vertical planes from airborne LiDAR point clouds, as well as its ability to efficiently utilize these planar primitives during model reconstruction.
Figure 9. Reconstructed building models from different datasets. (16) from the AHN3 dataset; (712) from the DALES dataset; (1316) from the Vaihingen dataset.
The quantitative evaluation results for the reconstruction models of each building in Figure 9 are shown in Table 2. The results show that the proposed method achieves high reconstruction accuracy across all test buildings, with RMSE values ranging from 0.06 m to 0.25 m. This improvement can be attributed to the proposed feature line constraint term, which effectively guides the selection of model planes, thereby enhancing reconstruction accuracy and enabling finer model details to better fit the input point clouds. In terms of reconstruction efficiency, the proposed method also performs well. Except for a few buildings with extremely complex roof structures, the reconstruction time for most building models is kept under 1 min. Furthermore, compared to traditional dense triangular mesh reconstruction methods [53], the proposed method generates simplified polygonal models with fewer faces and more compact topology, making it more suitable for large-scale urban modeling applications.
Table 2. Quantitative Evaluation Results of Reconstruction Models for Buildings in Figure 9. The report includes the number of input point clouds for each building, the number of faces in the reconstructed model, the fitting error, and the running time.
To further validate the effectiveness of the proposed reconstruction method, comparative experiments were conducted on three datasets—AHN3, DALES, and Vaihingen—against two state-of-the-art building reconstruction methods: City3D [28] and Geoflow [29]. Visual comparison results of some reconstructed models are shown in Figure 10. The comparison reveals that the model generated by our method demonstrates the highest structural integrity and regularity. In contrast, City3D’s reconstruction results exhibit deficiencies in detail recovery and are prone to topological errors in areas where multiple roof surfaces intersect. Geoflow’s reconstruction outcomes are significantly influenced by the quality of the input point cloud, struggling to reconstruct regular roof structures when the point cloud is sparse.
Figure 10. Comparison of Reconstruction Results from Different Methods: (a) Input point cloud; (b) Reconstruction result from the proposed method; (c) Reconstruction result from the City3D method; (d) Reconstruction result from the Geoflow method; (e) Reference reconstruction result.
Table 3 presents quantitative comparisons of the three methods across metrics, including model polygon count, reconstruction time, and RMSE. It is evident that while Geoflow demonstrates certain advantages in terms of polygon count and runtime, its reconstruction accuracy is significantly lower than that of the other two methods. The proposed method exhibits comparable computational efficiency to City3D, yet generates models with fewer polygons and more compact structures, achieving the highest reconstruction accuracy among the three approaches.
Table 3. Quantitative Evaluation Statistics for Reconstruction Comparison Methods. The report includes the total number of faces, average error, and running time.
Overall, both cross-dataset experimental results and multi-method comparative analyses demonstrate that the proposed method exhibits comprehensive advantages in terms of structural integrity, topological accuracy, and model compactness. It provides robust and reliable technical support for 3D reconstruction of urban buildings based on airborne LiDAR point clouds.

4.4. Discussion

As demonstrated by the aforementioned experimental results, when the roof point cloud is relatively complete, the proposed method can generate structurally sound and topologically consistent building models. However, in airborne LiDAR data, roof point clouds often exhibit various forms of missing data due to factors such as tree occlusion, shielding by adjacent high-rise buildings, or limited observation angles—a common occurrence in complex urban environments.
As illustrated in Figure 11a, missing data in the roof point cloud primarily occurs along building boundaries, resulting in incomplete roof outline information. Since our method relies on the roof outline to infer the overall building structure, damage to the outline makes it difficult to accurately reconstruct the building’s entire boundary, leading to missing data in localized areas of the reconstructed model.
Figure 11. Reconstruction Results and Topological Error Examples in Different Roof-Missing Scenarios. (a) Missing roof boundary points; (b) Cavities inside the roof; (c) Ring-shaped building.
For cases with cavities inside the roof (Figure 11b), since the outer contour remains intact and the roof plane extends and intersects in space within the “assume-select” framework, such cavities can be partially filled. However, the filled results may deviate significantly from the actual structure. Furthermore, when the input point cloud represents a ring-shaped building (Figure 11c), the extended roof plane may fill what should be an internal void, introducing topological errors.
The reconstruction framework presented in this paper utilizes only planar primitives, which suffices for the reconstruction needs of most urban buildings. However, in real-world scenarios, buildings with curved surfaces still exist, such as those indicated by the red rectangular box in Figure 12. Under this framework, such structures can only be approximated as segmented planar geometries.
Figure 12. Example of Segmented Planar Approximation Reconstruction Results for Curved Surfaces. (a) Input point cloud; (b) Reconstructed model.
Finally, it should be noted that the method described in this paper focuses primarily on reconstructing roof structures. Vertical facades are inferred from the three-dimensional points of the roof, and their positions may not perfectly align with the actual walls. As shown in Figure 13, for buildings with eaves, the reconstructed exterior walls of the model exhibit a certain degree of deviation from their true locations.
Figure 13. Example of exterior wall position deviation caused by eave structure. (a) Input point cloud; (b) Reconstructed model.

5. Conclusions

This paper proposes a method for generating closed, two-dimensional manifold-based three-dimensional polygonal building surface models from airborne LiDAR point clouds of individual structures. Overall, the main contributions of this work are threefold: (1) A planar segmentation method combining adaptive hypervoxels with graph cut optimization is introduced, which maintains both segmentation accuracy and planar integrity even under point cloud density variations; (2) It constructs a vertical plane inference strategy based on point cloud projection contours and height change detection, effectively completing the vertically oriented wall structures commonly missing in airborne LiDAR data; (3) It introduces feature line constraints within a “hypothesis-selection” reconstruction framework, enabling structured prior guidance for candidate plane selection. Experimental results on three publicly available airborne LiDAR point cloud datasets demonstrate that the proposed method outperforms existing approaches in model integrity, structural regularity, and reconstruction accuracy.
However, the proposed method still has room for improvement: when significant gaps exist in the roof point cloud, the completeness and topological correctness of the reconstruction results remain compromised. Additionally, the current reconstruction framework primarily relies on planar primitives, limiting its expressive power for surfaces or free-form structures. Future work will explore utilizing deep learning methods to complete missing roof point clouds and further extend the reconstruction primitives to include curved or hybrid primitive representations. This will enhance the method’s applicability and expressiveness in complex urban environments.

Author Contributions

Conceptualization, Zhao Hai; methodology, Cailin Li; software, Zhao Hai; validation, Baoyun Guo and Xianlong Wei; formal analysis, Zhuo Yang; investigation, Jinhui Zheng; writing—original draft preparation, Zhao Hai; writing—review and editing, Cailin Li; visualization, Zhao Hai. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Shandong Provincial Natural Science Foundation under Grant [No. ZR2022MD039].

Data Availability Statement

The airborne LiDAR datasets used in this study are publicly available. Detailed descriptions and access information are provided in the cited references.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Wang, R.; Peethambaran, J.; Chen, D. Lidar point clouds to 3-D urban models: A review. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 606–627. [Google Scholar] [CrossRef]
  2. Musialski, P.; Wonka, P.; Aliaga, D.G.; Wimmer, M.; Van Gool, L.; Purgathofer, W. A survey of urban reconstruction. In Computer Graphics Forum; Wiley Online Library: Hoboken, NJ, USA, 2013; pp. 146–177. [Google Scholar]
  3. Biljecki, F.; Stoter, J.; Ledoux, H.; Zlatanova, S.; Çöltekin, A. Applications of 3D City Models: State of the Art Review. ISPRS Int. J. Geo-Inf. 2015, 4, 2842–2889. [Google Scholar] [CrossRef]
  4. Cappelle, C.; El Najjar, M.E.; Charpillet, F.; Pomorski, D. Virtual 3D city model for navigation in urban areas. J. Intell. Robot. Syst. 2012, 66, 377–399. [Google Scholar] [CrossRef]
  5. Deren, L.; Wenbo, Y.; Zhenfeng, S. Smart city based on digital twins. Comput. Urban Sci. 2021, 1, 4. [Google Scholar] [CrossRef]
  6. Yu, M.; Yang, C.; Li, Y. Big Data in Natural Disaster Management: A Review. Geosciences 2018, 8, 165. [Google Scholar] [CrossRef]
  7. Waqas, H.; Jiang, Y.H.; Shang, J.G.; Munir, I.; Khan, F.U. An Integrated Approach for 3D Solar Potential Assessment at the City Scale. Remote Sens. 2023, 15, 5616. [Google Scholar] [CrossRef]
  8. Miller, H.; Tolle, K. Big Data for Healthy Cities: Using Location-Aware Technologies, Open Data and 3D Urban Models to Design Healthier Built Environments. Built Environ. 2016, 42, 441–456. [Google Scholar] [CrossRef]
  9. Bauchet, J.-P.; Sulzer, R.; Lafarge, F.; Tarabalka, Y. Simplicity: Reconstructing buildings with simple regularized 3D models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 17–18 June 2024; pp. 7616–7626. [Google Scholar]
  10. Bauchet, J.-P.; Lafarge, F. City reconstruction from airborne LiDAR: A computational geometry approach. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2019, 4, 19–26. [Google Scholar] [CrossRef]
  11. Li, M.; Rottensteiner, F.; Heipke, C. Modelling of buildings from aerial LiDAR point clouds using TINs and label maps. ISPRS J. Photogramm. Remote Sens. 2019, 154, 127–138. [Google Scholar] [CrossRef]
  12. Ledoux, H.; Biljecki, F.; Dukai, B.; Kumar, K.; Peters, R.; Stoter, J.; Commandeur, T. 3dfier: Automatic reconstruction of 3D city models. J. Open Source Softw. 2021, 6, 2866. [Google Scholar] [CrossRef]
  13. Zhou, Q.-Y. 3D Urban Modeling from City-Scale Aerial LiDAR Data. Ph.D. Thesis, University of Southern California, Los Angeles, CA, USA, 2012. [Google Scholar]
  14. Rottensteiner, F.; Sohn, G.; Gerke, M.; Wegner, J.D.; Breitkopf, U.; Jung, J. Results of the ISPRS benchmark on urban object detection and 3D building reconstruction. ISPRS J. Photogramm. Remote Sens. 2014, 93, 256–271. [Google Scholar] [CrossRef]
  15. Nan, L.; Wonka, P. Polyfit: Polygonal surface reconstruction from point clouds. In Proceedings of the IEEE international Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2353–2361. [Google Scholar]
  16. Overby, J.; Bodum, L.; Kjems, E.; Iisoe, P. Automatic 3D building reconstruction from airborne laser scanning and cadastral data using Hough transform. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2004, 34, 1–6. [Google Scholar]
  17. Fischler, M.A.; Bolles, R.C. Random sample consensus: A paradigm for model fitting with applications to image analysis and automated cartography. Commun. ACM 1981, 24, 381–395. [Google Scholar] [CrossRef]
  18. Zuliani, M.; Kenney, C.S.; Manjunath, B. The multiransac algorithm and its application to detect planar homographies. In Proceedings of the IEEE International Conference on Image Processing 2005, Genova, Italy, 11–14 September 2005; pp. III–153–III–156. [Google Scholar]
  19. Schnabel, R.; Wahl, R.; Klein, R. Efficient RANSAC for point-cloud shape detection. In Computer Graphics Forum; Wiley Online Library: Hoboken, NJ, USA, 2007; pp. 214–226. [Google Scholar]
  20. Lafarge, F.; Mallet, C. Creating large-scale city models from 3D-point clouds: A robust approach with hybrid representation. Int. J. Comput. Vis. 2012, 99, 69–85. [Google Scholar] [CrossRef]
  21. Rabbani, T.; Van Den Heuvel, F.; Vosselmann, G. Segmentation of point clouds using smoothness constraint. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2006, 36, 248–253. [Google Scholar]
  22. Sun, S.; Salvaggio, C. Aerial 3D building detection and modeling from airborne LiDAR point clouds. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2013, 6, 1440–1449. [Google Scholar] [CrossRef]
  23. Zhou, Q.-Y.; Neumann, U. 2.5 D Dual Contouring: A Robust Approach to Creating Building Models from Aerial LiDAR Point Clouds. In Proceedings of the European Conference on Computer Vision, Crete, Greece, 5–11 September 2010; pp. 115–128. [Google Scholar]
  24. Zhou, Q.-Y.; Neumann, U. 2.5 D building modeling by discovering global regularities. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 326–333. [Google Scholar]
  25. Sampath, A.; Shan, J. Segmentation and reconstruction of polyhedral building roofs from aerial lidar point clouds. IEEE Trans. Geosci. Remote Sens. 2009, 48, 1554–1567. [Google Scholar] [CrossRef]
  26. Chen, D.; Wang, R.; Peethambaran, J. Topologically aware building rooftop reconstruction from airborne laser scanning point clouds. IEEE Trans. Geosci. Remote Sens. 2017, 55, 7032–7052. [Google Scholar] [CrossRef]
  27. Xie, L.; Hu, H.; Zhu, Q.; Li, X.; Tang, S.; Li, Y.; Guo, R.; Zhang, Y.; Wang, W. Combined rule-based and hypothesis-based method for building model reconstruction from photogrammetric point clouds. Remote Sens. 2021, 13, 1107. [Google Scholar] [CrossRef]
  28. Huang, J.; Stoter, J.; Peters, R.; Nan, L. City3D: Large-scale building reconstruction from airborne LiDAR point clouds. Remote Sens. 2022, 14, 2254. [Google Scholar] [CrossRef]
  29. Peters, R.; Dukai, B.; Vitalis, S.; van Liempt, J.; Stoter, J. Automated 3D reconstruction of LoD2 and LoD1 models for all 10 million buildings of the Netherlands. Photogramm. Eng. Remote Sens. 2022, 88, 165–170. [Google Scholar] [CrossRef]
  30. Nobert, H. Virtural city models from laser altimeter and 2D map data. Photogramm. Eng. Remote Sens. 1999, 65, 787–795. [Google Scholar]
  31. Henn, A.; Gröger, G.; Stroh, V.; Plümer, L. Model driven reconstruction of roofs from sparse LIDAR point clouds. ISPRS J. Photogramm. Remote Sens. 2013, 76, 17–29. [Google Scholar] [CrossRef]
  32. Li, M.; Wonka, P.; Nan, L. Manhattan-world urban reconstruction from point clouds. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2016; pp. 54–69. [Google Scholar]
  33. Ikehata, S.; Yang, H.; Furukawa, Y. Structured indoor modeling. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1323–1331. [Google Scholar]
  34. Coughlan, J.; Yuille, A.L. The manhattan world assumption: Regularities in scene statistics which enable bayesian inference. Adv. Neural Inf. Process. Syst. 2000, 13, 417–424. [Google Scholar]
  35. Zhang, W.; Li, Z.; Shan, J. Optimal model fitting for building reconstruction from point clouds. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 9636–9650. [Google Scholar] [CrossRef]
  36. Li, Z.; Shan, J. RANSAC-based multi primitive building reconstruction from 3D point clouds. ISPRS J. Photogramm. Remote Sens. 2022, 185, 247–260. [Google Scholar] [CrossRef]
  37. Lin, H.; Gao, J.; Zhou, Y.; Lu, G.; Ye, M.; Zhang, C.; Liu, L.; Yang, R. Semantic decomposition and reconstruction of residential scenes from LiDAR data. ACM Trans. Graph. (TOG) 2013, 32, 1–10. [Google Scholar] [CrossRef]
  38. Bizjak, M.; Mongus, D.; Žalik, B.; Lukač, N. Novel half-spaces based 3D building reconstruction using airborne LiDAR data. Remote Sens. 2023, 15, 1269. [Google Scholar] [CrossRef]
  39. Xiong, B.; Jancosek, M.; Oude Elberink, S.; Vosselman, G. Flexible building primitives for 3D building modeling. ISPRS J. Photogramm. Remote Sens. 2015, 101, 275–290. [Google Scholar] [CrossRef]
  40. Xu, B.; Jiang, W.; Li, L. HRTT: A hierarchical roof topology structure for robust building roof reconstruction from point clouds. Remote Sens. 2017, 9, 354. [Google Scholar] [CrossRef]
  41. Hu, P.; Yang, B.; Dong, Z.; Yuan, P.; Huang, R.; Fan, H.; Sun, X. Towards reconstructing 3D buildings from ALS data based on gestalt laws. Remote Sens. 2018, 10, 1127. [Google Scholar] [CrossRef]
  42. Chen, Z.; Ledoux, H.; Khademi, S.; Nan, L. Reconstructing compact building models from point clouds using deep implicit fields. ISPRS J. Photogramm. Remote Sens. 2022, 194, 58–73. [Google Scholar] [CrossRef]
  43. Chen, Z.; Shi, Y.; Nan, L.; Xiong, Z.; Zhu, X.X. PolyGNN: Polyhedron-based graph neural network for 3D building reconstruction from point clouds. ISPRS J. Photogramm. Remote Sens. 2024, 218, 693–706. [Google Scholar] [CrossRef]
  44. Liu, Y.; Obukhov, A.; Wegner, J.D.; Schindler, K. Point2Building: Reconstructing buildings from airborne LiDAR point clouds. ISPRS J. Photogramm. Remote Sens. 2024, 215, 351–368. [Google Scholar] [CrossRef]
  45. Li, L.; Song, N.; Sun, F.; Liu, X.; Wang, R.; Yao, J.; Cao, S. Point2Roof: End-to-end 3D building roof modeling from airborne LiDAR point clouds. ISPRS J. Photogramm. Remote Sens. 2022, 193, 17–28. [Google Scholar] [CrossRef]
  46. Jiang, T.; Wang, Y.; Zhang, Z.; Liu, S.; Dai, L.; Yang, Y.; Jin, X.; Zeng, W. Extracting 3-D structural lines of building from ALS point clouds using graph neural network embedded with corner information. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–28. [Google Scholar] [CrossRef]
  47. Lin, Y.; Wang, C.; Zhai, D.; Li, W.; Li, J. Toward better boundary preserved supervoxel segmentation for 3D point clouds. ISPRS J. Photogramm. Remote Sens. 2018, 143, 39–47. [Google Scholar] [CrossRef]
  48. Actueel Hoogtebestand Nederland (AHN). AHN dataset. 2018. Available online: https://www.ahn.nl (accessed on 5 March 2025).
  49. Varney, N.; Asari, V.K.; Graehling, Q. DALES: A large-scale aerial LiDAR data set for semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 13–19 June 2020; pp. 186–187. [Google Scholar]
  50. Rottensteiner, F.; Sohn, G.; Jung, J.; Gerke, M.; Baillard, C.; Benitez, S.; Breitkopf, U. The ISPRS benchmark on urban object classification and 3D building reconstruction. ISPRS Ann. Photogramm. Remote Sens. Spatial Inf. Sci. 2012, I-3, 293–298. [Google Scholar] [CrossRef]
  51. Rutzinger, M.; Rottensteiner, F.; Pfeifer, N. A comparison of evaluation techniques for building extraction from airborne laser scanning. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2009, 2, 11–20. [Google Scholar] [CrossRef]
  52. Yan, J.; Shan, J.; Jiang, W. A global optimization approach to roof segmentation from airborne lidar point clouds. ISPRS J. Photogramm. Remote Sens. 2014, 94, 183–193. [Google Scholar] [CrossRef]
  53. Verdie, Y.; Lafarge, F.; Alliez, P. LOD generation for urban scenes. ACM Trans. Graph. 2015, 34, 15. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.