Next Article in Journal
Properties of Concrete Influenced by Plastic Materials
Previous Article in Journal
Citywalk in Chinese Metropolises: A Multidimensional Framework for Evaluating Urban Walking Environments
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Scan-to-EDTs: Automated Generation of Energy Digital Twins from 3D Point Clouds

by
Oscar Roman
1,2,
Maarten Bassier
3,
Giorgio Agugiaro
4,*,
Ken Arroyo Ohori
4,
Elisa Mariarosaria Farella
2 and
Fabio Remondino
2
1
EICS and DII Department, University of Trento, via Sommarive 9, 38123 Trento, Italy
2
3DOM 3D Optical Metrology, Bruno Kessler Foundation, via Sommarive 18, 38123 Trento, Italy
3
Department of Civil Engineering, TC Construction-Geomatics, KU Leuven-Faculty of Engineering Technology, 9000 Ghent, Belgium
4
3D Geoinformation Group, Department Urbanism, Faculty of Architecture and Built Environment, Delft University of Technology, 2628 BL Delft, The Netherlands
*
Author to whom correspondence should be addressed.
Buildings 2025, 15(22), 4060; https://doi.org/10.3390/buildings15224060
Submission received: 26 September 2025 / Revised: 4 November 2025 / Accepted: 8 November 2025 / Published: 11 November 2025
(This article belongs to the Section Construction Management, and Computers & Digitization)

Abstract

Digital Twins (DTs) are transforming construction and energy management sectors by integrating 3D surveying, monitoring, Building Performance Simulation (BPS), and Building Energy Simulation (BES) from the earliest design or retrofit stages. Moreover, dynamic thermal simulations further support energy performance assessments by modeling indoor conditions to meet comfort and efficiency targets. However, their reliability depends on accurate, standards-compliant 3D building models, which are costly to create. This research introduces a complete framework for automatically generating energy-focused Digital Twins (EDTs) directly from unstructured point clouds. Combining Deep Learning-based instance detection, Scan-to-BIM techniques, and computational geometry, the method produces simulation-ready models without manual intervention. The resulting EDTs streamline early-stage performance evaluation, enable scenario testing, and enhance decision making for energy-efficient retrofits, advancing smart-building design through predictive simulation.

1. Introduction

Achieving the EU’s 2050 zero-emission building target necessitates immediate, cross-sectoral intervention, as building operations globally accounted for 30% of final energy consumption and 26% of energy-related emissions in 2022, including 8% from direct sources and 18% from electricity and heat production [1]. The EU’s estimated 24 billion m2 of floor area, 75% residential and 25% non-residential, consumes 64.4% of residential energy for heating and cooling [2,3], with potential reductions of up to 70% through improved insulation, heat pumps and solar integration [4].
Achieving this reduction at a large scale requires digitization across the AECOO (Architecture, Engineering, Construction, Owners and Operations) sector, where recent growth in tools supporting interoperability between Building Information Modeling (BIM) and energy simulation platforms reflects the industry’s shift toward data-driven energy optimization [5]. By automating the generation of Energy Digital Twins (EDTs) from raw spatial data, the proposed framework directly speeds up large-scale energy audits and retrofit planning, thereby contributing to the EU’s 2050 decarbonization targets by reducing modeling costs, retrofit analysis, and enabling scalable, simulation-driven optimization. Furthermore, it is well suited for building-stock mapping and for creating a complete as-is digital record of each building at different scales.

1.1. Context of the DT for Building Applications

Digital Twins (DTs) are increasingly important in building operations, integrating geometric models, real-time Digital Shadows (DSs) [6], and actuators for closed loop control to form a comprehensive virtual representation of physical systems.
In common approaches (Figure 1), the DS is the continuous stream of time-stamped measurements and events from the asset, whereas the geometric Digital Twin (gDT) is a structured 3D representation (e.g., solid, Boundary Representation (B-Rep), Constructive Solid Geometry (CSG), or Building Information Modeling (BIM)) that evolves slowly and serves as the authoritative geometric–topological–semantic index. Figure 1 shows a reference architecture widely used in the literature: the DS mirrors the physical state, while the gDT provides the spatial index from which domain-specific models, such as the structural Finite Element Model (FEM), Discrete Element Method (DEM), and Building Management System (BMS) models, are derived. The DS does not store geometry; instead, it references gDT elements via stable identifiers or coordinates.
In this research, DS data are instead embedded within the gDT schema, producing a unified representation that couples geometry with historical sensor data. A DT then links the gDT, these domain models, and the DS in a bi-directional loop for state estimation, simulation, and control. Artificial Intelligence (AI), Machine Learning (ML), Deep Learning (DL), and Physics-Informed Neural Networks (PINNs) elevate the twin to an intelligent, autonomous system capable of forecasting and self-optimization. Multiple twins can be joined to form a Federated Digital Twin (FDT) for system-level optimization. Integration with the BMS enables real-time monitoring [7,8,9], predictive maintenance [10,11,12,13], and energy optimization [14,15], further enhanced by PINN-driven Data-Driven Decision Support Systems (DDSS) [16].
However, current tools often suffer from fragmented interoperability, incompatible file formats, divergent model structures, and inconsistent communication protocols, largely due to a lack of standardization [17]. These issues cause data loss, workflow inefficiencies, and reduced analytical coherence [18,19]. Addressing these challenges requires semantic integration frameworks that unify heterogeneous data sources into high-fidelity, queryable models. Structured hierarchically and enriched with spatial reasoning and semantics, such models ground intelligent systems and support accurate, scalable, performance-oriented decision making across the building life cycle.

1.2. Aim of the Research

This research presents a holistic end-to-end framework that generates a multi-level geometric Digital Twin from surveyed point clouds, unifying (i) solid and (ii) surface B-Reps with consistent topology, in particular, watertight and 2-manifold envelopes that preserve face-edge-vertex incidence and validated zone-surface adjacencies across building, thermal-zone, and surface levels, enriched with semantic, topological, and energy data for advanced simulation within a single graph model that supports real-time IoT connection and performance optimization. Unlike prior Scan-to-BIM or DT creation approaches that rely on manual modeling or intermediate BIM translation, the proposed unified data-driven pipeline produces energy-compliant DTs directly from unstructured point clouds. The principal novelties are (i) integration of topology and energy semantics within a single model; (ii) elimination of traditional BIM steps; and (iii) complete automation from geometry to energy modeling.
The framework advances state-of-the-art DT generation by the following:
(i)
Deriving simulation-ready Energy Digital Twins (EDTs) directly from point clouds without manual steps, BIM stages, or reliance on proprietary authoring tools;
(ii)
Delivering watertight, simulation-ready models with full topology, semantics, and energy attributes for reasoning, querying and interoperability;
(iii)
Unifying solid and B-Rep representations in a single multi-level gDT and integrated graph for simulations, monitoring, IoT linking, and cross-scale navigation;
(iv)
Enabling standards-based exchange via gbXML and optional IFC to ensure interoperability while avoiding redundant steps and minimizing data loss.

2. State of the Art and Related Research

Systematic reviews [20,21,22] indicate that DTs for building energy applications are typically developed through complex, multi-phase workflows that are still evolving from conventional BIM and Building Energy Modeling (BEM) practices. Conventional workflows start with manual BIM stages, then assemble a federated BEM with Mechanical, Electrical, and Plumbing (MEP) systems. From this, an energy-analytical model is derived by defining thermal zones, constructions, loads, and schedules, and the BEM is then exported to IFC or gbXML for simulation and visualization. Monitoring and sensor integration are added later to reach a DT for interactive control and energy optimization.
Despite this progress, most DT implementations remain partially static and fragmented. Many DTs reported in the literature [23] are more accurately classified as geometric Digital Twins (gDTs), as they lack a continuous, bidirectional connection to IoT sensors. When data flow occurs only in one direction, from the physical asset to its digital representation, the model is better defined as a Digital Shadow (DS). In both cases, the absence of live, two-way communication means the system cannot respond dynamically to real-world changes, which is a defining feature of a fully realized DT.
Although the literature [23] often implies a direct BIM-to-DT link, the relationship is more complex and the connection is neither linear nor efficient. BIM models, whether manual or Scan-to-BIM generated with classification, segmentation, and reconstruction, are static, solid, and design oriented: they aid coordination and documentation but are ill-suited to lifecycle asset management and integrating heterogeneous data in BIM lowers usability and raises system complexity (Figure 2). For this reason, integrating real-time data often requires specialized tools or platforms and alternative model types. Tools like Dynamo [24] for Revit (version 2.17.0.3472, Autodesk, San Rafael, CA, USA) now support continuous and historical data flows, while platforms like Autodesk Forge enable cloud-based access and visualization, both essential for operational DTs.
Consequently, there is growing research interest in bypassing traditional BIM-based workflows by directly generating energy-oriented geometric Digital Twins (EDTs) from unstructured 3D data. The innovation of the present study lies in addressing this gap by bypassing BIM where possible and generating energy-oriented EDTs directly from raw point clouds. The following sections review the state of the art across the full workflow, from point cloud segmentation and Scan-to-BIM geometry to the generation of EDTs and a gDT ready to be enriched with real-time and historical data for monitoring and simulation.

2.1. Advancements in Semantic Segmentation: Transformers and 3D Open Vocabularies

In recent years, a wide range of point cloud classification and semantic segmentation methods have emerged, broadly categorized into image-based and point-based approaches [25]. Given the unstructured nature of 3D point clouds, traditional Convolutional Neural Networks (CNNs) have been adapted for effective point cloud processing. In segmentation, both CNNs and Transformers have evolved [26], incorporating self-attention mechanisms [27], channel-wise attention, and graph-based structures for improved sensitivity [28,29]. Hierarchical global feature aggregation [30], Point Cloud Local Auxiliary Blocks (PLABs) for neighborhood enhancement [31], edge-aware modules for fine detail [32], and multi-scale local aggregation strategies further advance performance. Kernel methods like KPConvX [33] use deformable kernels, while the Point Patch Transformer (PPT) [34], combining Points2Patches and positional encoding to capture local and global contexts for state-of-the-art performance. Emerging vocabulary free approaches [35,36] use vision-language models, and spectral clustering to merge Superpoints, enabling category-agnostic object discovery from raw 3D data [37].

2.2. Image-Based Classification and Semantic Segmentation

Semantic segmentation has also advanced rapidly in image segmentation. Key advances include the Segment Anything Model (SAM) for prompt-based, zero-shot segmentation with broad generalization [38] and GroundingDINOv2 using self-supervised, transferable features for stronger dense prediction [39]. These advances highlight self-supervised learning and transformer architectures in vision, enabling segmentation without task-specific training. Among these, K-Net [40] enhances object grouping using Transformers for better global context, while Pyramid Fusion Transformer (PFT) [41] leverages multi-scale features for top-tier performance. Models like MaskFormer [42] and Graph-Segmenter [43] further improve segmentation with boundary-aware attention and graph-based refinement.

2.3. Object Detection in Challenging Environments: Latest Strategies and Techniques

Detecting small or occluded indoor objects is challenging, especially when they are camouflaged within their surroundings, such as wall-integrated radiators, recessed lights, or embedded sensors [44]. In such settings, the goal often shifts from exact localization to detecting object presence within a Region of Interest (ROI). Region-based Convolutional Neural Networks (R-CNNs) support this by proposing candidate regions for detection and classification [45]. Multi-scale feature extraction addresses varying object sizes, while contextual recognition uses spatial cues to infer object presence. Region Proposal Networks (RPNs) suggest object locations and bounding box-based loss functions like Intersection over Union (IoU) evaluate object presence. Anchor boxes help define object sizes and shapes, and probabilistic models assign confidence scores, compensating for occlusion and partial visibility. Similar approaches are used in complex domains like forestry [46] and harvesting [47]. Finally, a promising 3D Large Language Model (LLM) has been introduced in [48], capable of processing point cloud data to generate structured 3D scene understanding for complex analysis.

2.4. Late Fusion Approaches

Late fusion is a multi-modal integration strategy where features from different data sources are processed separately and combined at the decision-making stage [49]. This preserves the unique strengths of each modality and minimizes redundancy. It is widely used in applications involving heterogeneous sensors, such as RGB cameras, LiDAR, and Global Navigation Satellite System/Inertial Measurement Unit (GNSS/IMU) [50], to enhance robustness in challenging environments [51] like tunnels [52], cluttered indoor spaces, or dense vegetation [53]. In urban-scale building segmentation, late fusion merges high-resolution satellite imagery [54], Digital Surface Models (DSMs), and ancillary data (e.g., OpenStreetMap), improving instance-level accuracy and boundary delineation. Similarly, in 3D reality capture, late fusion of features from images and point clouds across parallel networks enhances both object detection and geometric fidelity by leveraging complementary spatial and semantic cues.

2.5. Advancements in Scan-to-BIM

Automated BIM generation remains an emerging field with limited comprehensive methods, marking it as a key area of ongoing research [55]. Successful Scan-to-BIM approaches typically rely on structured classification and segmentation of raw point cloud data and, in general, they follow three main stages [56]: (i) point cloud segmentation, (ii) classification and instance segmentation, and (iii) 3D modeling and evaluation. These are essential to create BIMs of existing buildings and support applications such as indoor navigation [57], construction robotics, and Simultaneous Localization and Mapping (SLAM) technologies in the AEC sector [58].
For example, SLABIM [59] merges BIM with SLAM sensor data to enhance registration, localization, and semantic mapping in complex indoor spaces. These stages also support change detection, structural deformation monitoring [60], and quality control [61,62,63]. The use of Terrestrial and Mobile Laser Scanning (TLS or MLS) enables detailed 3D point cloud capture, despite challenges like clutter, occlusion, and reflective or transparent surfaces [64]. Recent advances include computational geometry-based methods [65], graph-based methods [66], and Reversible Jump Markov Chain Monte Carlo (RJMCMC) techniques [67] for improved 3D reconstruction. These developments aid in building semantic-rich DT for structural and energy analysis.

2.6. BIM, BEM, and Building Management Systems (BMSs)

Building Energy Models (BEMs) are increasingly critical for energy-efficient design and retrofitting, complementing BIM’s strengths in 3D visualization and construction planning. BIM–BEM integration [68,69,70] enables comprehensive model generation and accurate energy analysis, yet faces challenges like interoperability, between solid and B-Rep representation and the need for expert input. Integrating MEP systems enhances detail and enables clash detection to avoid on-site conflicts. Advances in object detection [71], extended reality [72], and ML/DL methods [15,73] support energy optimization, smart modeling [74,75], and system management via graph-based approaches [76] and HVAC optimization [77]. However, since BIM is mostly a static model, it remains underutilized in facility management, particularly during the operational phase, which accounts for over 80% of lifecycle costs. Detailed simulations at the building level, with progressive data enrichment and semantic alignment with BEM schemas (e.g., EnergyPlus .IDF or .epJSON), are supported by EnergyPlus [14], while tools like Ladybug and HoneyBee [78] extend analysis to city-wide energy modeling, district consumption, and demand forecasting [79,80], enabling data-driven urban planning.

2.7. Topologic BIM (TBIMs)

Scan-to-BIM is a complex first step toward creating DTs for building applications. Bridging BIM’s solid and heavy geometry and product-focused outputs with space-oriented simulation workflows requires semantic consistency, spatial reasoning, and topological connectivity. The TBIM methodology [81] addresses this by modeling buildings with topological abstractions, such as rooms as cells, partitions as faces, and clusters as complex cells, enriched via rule-based modeling and topological queries to assign energy properties [82,83]. TBIM reduces geometric complexity [84] while preserving semantic richness, enabling compatibility with Building Performance Simulation (BPS) tools [85]. It supports non-manifold topology, hierarchical relationships, geometry-agnostic modeling, semantic metadata, and interoperability. Rather than adhering to a single standard, TBIMs are purpose-driven representations that rely on various building ontologies, such as IFC, Building Object Typology (BOT), or Brick, depending on the target application, be it compliance checking, energy simulation, or real-time DT integration.

2.8. Digital Twins for Buildings

The DT concept originated with NASA’s Apollo 13 for real-time monitoring [86]. Currently, DTs are used in buildings, industry, manufacturing, and cities for management, maintenance, and performance optimization [87,88]. Unlike static BIM, DTs fuse real time and historical data from the physical asset and its virtual model to inform structural, energy, and operational decisions [89]. In energy applications, they improve fault detection, anomaly prediction, and efficiency, and track indoor conditions such as temperature, CO2, and humidity for predictive maintenance [90]. Combined with AI, they run simulations that reveal inefficiencies and optimize performance in real time [3,6,8,9].
A key challenge is integrating geometry, semantics, and external datasets within a single real-time framework. Since DTs are visual and data-driven replicas of physical systems [91,92], seamless interoperability across diverse sources is essential for accurate monitoring, simulation, and decision making.
The development of DTs can be classified by functional maturity [22]:
  • DT Level 1: Real-time sensor integration for monitoring, visualization, and diagnostics.
  • DT Level 2: Dynamic simulation, forecasting, and optimization (also from historic data).
  • DT Level 3: Analytics for failure prediction and strategic building management, contingent on historical data availability.
A key distinction between a geometric DT and a full DT (Figure 3) is the integration of real-time data and the ability to simulate behavior from that data. This enables progressively advanced capabilities: monitoring (Level 1), simulation (Level 2), and predictive maintenance (Level 3) [22,93].
These capabilities align with different DT types: Geometric (static 3D models), Behavioral (system performance), Operational (real-time and historical data), and Predictive (future state forecasting).

3. Methodology

Figure 4 introduces the pipeline that generates energy-oriented geometric digital twins (EDTs) directly from raw point clouds, extending [94] into a structured, semantically enriched, multi-level DT for monitoring and preliminary simulations.
The framework builds a geometrically accurate, topologically coherent model enriched with energy metadata through the following phases:
  • Object Classification and Instance Segmentation (Section 3.1 and Section 3.2): Detect architectural components in the 3D point cloud, merging point-based and image-based outputs via late fusion [94].
  • Scan-to-BIM Process (Section 3.3): Build an intermediate solid model (Figure 4, step A) with a graph representation of geometry and spatial configuration, aligning with object-level reconstruction in [94] and solid mesh representation [66].
  • Scan-to-BEM Process (Section 3.4): Augment the model with energy devices, their localization, and their data for integration with building energy modeling workflows.
  • Topologic Model Generation (Section 3.5): Convert the 2D Topologic Map [95] into a 3D graph-based topology with adjacencies, containment hierarchies, and watertight room enclosures, and fully closed 3D spaces with seamlessly joined surfaces, ensuring validity for energy and spatial simulations.
  • Geometric Digital Twin output (Section 3.6): Derive the final multi-level gDT that integrates geometry, topology, and semantic energy attributes in a graph-based model, ready for simulation-driven DT development.
The system builds a progressively enriched graph with two hierarchical views:
  • Volumetric model (Figure 4, step B) that defines watertight room level enclosures for thermal zoning, a Topologic Solid Model (TSM);
  • Semantic model (Figure 4, step C) that encodes surfaces and thematic partitions with material and use metadata.
The resulting gDT (Figure 4, step D) supports hierarchical visualization, semantic queries, graph navigation, and real-time data integration. Semantic and material enrichment is achieved through Information Loading Dictionaries (ILDs), which add space-specific functional, regulatory, and energy data such as occupancy and device specifications. Coupled with IoT sensors and simulation engines, the static gDT is transformed into an operational Digital Twin.

3.1. Object Classification and Instance Segmentation: Point Clouds

For the captured point clouds, the methodology leverages the Pointcept framework that integrates PointTransformer version 3 [26,96] for classification tasks and the Segmentor module for instance segmentation, ensuring identification of building elements. During the preprocessing phase, point cloud data is subsampled to a 1 cm resolution using a Voxel Grid Filtering (VGF) approach and partitioned into smaller grid-based subsets, enhancing processing efficiency and scalability. The model is trained to detect five distinct classes, encompassing both primary elements, such as ceilings, floors, and walls, and secondary elements, including doors and windows. As highlighted in [94], the detection phase exhibits robust performance in recognizing objects with distinct geometric characteristics, such as floors and walls.

3.2. Object Classification and Instance Segmentation: Images

To address detection challenges, image-based strategies are employed to identify columns [97] and openings [98] from the point cloud. The raw point cloud is (i) rasterized into a top view projection at 1 cm resolution and (ii) aligned with the global xy axes by estimating the building’s principal orientation from the oriented bounding box (OBB) of the entire building. A YOLO detector [97] is then applied to localize columns in the raster image. For openings, (i) virtual cameras are placed 1 m orthogonal to each wall face, from which (ii) high-resolution orthographic views are rasterized using wall-aligned octree meshes. GroundingDINO [39] is applied to these projections and dimensional constraints filter and classify door and window candidates. Finally, all detections are mapped back into the 3D world space as Oriented Bounding Boxes (OBBs) via inverse transformations from the projection frames. A late-fusion scheme (from Section 3.1 and Section 3.2) based on class-wise model selection was adopted. Let p c ( P T ) ( x )   denote per-class logits or probabilities from the Point Transformer on the 3D point cloud and p c ( I M G ) ( x ) those obtained by reprojecting image-based detections (YOLO, GroundingDINO) onto the point cloud. On a held-out validation set, per-class performance (mIoU) is estimated and a provider π c { PT ,   IMG } is chosen for each class c . At inference, the fused score s c x for class c is the following:
s c x = p c P T x , i f π c = P T p c I M G x , i f π c = I M G ŷ x = a r g m a x c s c x
where π c denotes the per-class provider selected on validation mIoU (ties broken by calibrated confidence). Image predictions are reprojected onto the point cloud to ensure spatial alignment. Spatial coherence is enforced via connected-component smoothing and majority voting within local 3D regions. Concretely, geometry-dominant classes are taken from the Point Transformer, whereas small or texture/appearance-dominant classes are sourced from YOLO/GroundingDINO. This achieves the benefits of both modalities without heuristic per-pixel weighting.

3.3. From Geometric Graph to Reconstructed Objects

The geometric model leverages the graph-structured framework [66] to extract class_id (the element category in the ontology, such as doors, walls, windows) and object_id (the unique instance identifier for the specific physical element) from the classified point cloud, systematically organizing all related information within this structure. This framework includes essential components, such as OBB and associated point data, along with georeferenced information, creating a robust foundation for spatial data representation. By extracting and processing these points [94], the various object classes are reconstructed. In the following subsections, we detail the reconstruction procedure for each class element.
Let 𝒟 represent the complete point cloud. Following the classification phase, 𝒟 can be described as consisting of the following components (Equation (2)):
D = U + F + C + W + L + O
where:
  • F represents the floor class, serving as the base to which the lower bonds of all walls are connected (Section 3.3.1);
  • C represents the ceiling class, serving as the upper connection point for the tops of all wall elements (Section 3.3.1);
  • W denotes the wall class, containing all identified vertical partition elements (Section 3.3.2);
  • L refers to the column class, capturing all structural columns within the environment;
  • O denote the sets of doors and windows, so it defines the generalized openings class (Section 3.3.3);
  • U represents the unclassified class, encompassing all elements that do not fall into any of the defined categories above.

3.3.1. Floors ( F ) and Ceiling ( C ) Class

The method robustly extracts horizontal structural elements, from floors ( F ) and ceilings ( C ) p o i n t   c l o u d   c l a s s e s , from the original 3D point cloud D , using geometric and spatial analysis. Surface normals are first estimated using local Principal Component Analysis (PCA) with a radius r =   0.30   m . This radius defines the PCA neighborhood, where smaller values (0.10–0.15 m) capture finer detail but are more noise sensitive, while larger ones (0.50–0.60 m) yield smoother slab normals yet risk over-smoothing and merging adjacent levels. Points satisfying n z 0.98 are classified as horizontally aligned candidates. These are clustered using the DBSCAN [98] algorithm ε = 0.5 m , minPts = 30 to group co-planar regions and suppress noise. Each cluster is projected onto the xy plane and bounded by a minimum Oriented Bounding Box (mOBB), capturing its extent and orientation. Floor–ceiling pairs are identified by evaluating vertical spacing between mOBB centroids Δ z 2.0 m and enforcing a horizontal overlap threshold (IoU or projection overlap ≥ 85%). This stratifies the point cloud into building levels, forming the basis for semantic segmentation and volumetric reasoning.
Finally, for each pair z max ,   z min   a n d   z mean (Equation (3)) are stored to define vertical extents and ensure geometric connectivity between floors, ceilings, and adjoining walls.
z mean = z min + z max z min 2 .

3.3.2. Wall Class ( W )

The topology of walls is established through a robust workflow inspired by the methodology in [58] comprising an initial noise-filtering stage followed by a Random Sample Consensus (RANSAC)-based plane segmentation. For each wall cluster W = {wi, …, wk}, a dominant supporting plane is identified. This plane is characterized by a standard planar equation (Equation (4)) of the following form:
a x + b y + c z + d = 0
where the coefficients a, b, c, d are the unit normal vector of the wall plane and d is the signed distance from the origin. The coefficients are obtained by minimizing the least-squares error function (Equation (5))
E = i = 1 N a x i + b y i + c z i + d 2 a 2 + b 2 + c 2
which corresponds to minimizing the sum of squared orthogonal distances between the plane and the observed points. Walls with a thickness below a specified threshold (tth = 0.12 m) were excluded from reconstruction, consistent with common dimensions of interior partitions in standard construction practice. The remaining walls are grouped into clusters based on two criteria: proximity and orientation similarity. For the proximity criterion, the distance between two clusters 𝒲1 and 𝒲2 is computed (Equation (6)), where d W 1 , W 2 t d , with td as the distance threshold.
d W 1 , W 2 = min w 1 i W 1 , w 2 j W 2 | w 1 i w 2 j |
On the other hand, the orientation criterion evaluates the angular similarity between clusters (Equation (7)), and they are merged only when both the distance and orientation criteria are simultaneously satisfied.
θ W 1 , W 2 = arccos n W 1 n W 2 | n W 1 | | n W 2 |
To analyze connections between walls, the algorithm examines intersections and spatial relationships by leveraging geometric features and computational methods. Each wall is represented by its start and end points   p start , p end , its normal vector ( n ) , and orthogonal offsets p start , p end   that define the wall’s boundaries in 3D space. The orthogonal offsets are calculated as (Equation (8)):
p start = p start + t i 2 n , p end = p end + t i 2 n
where t i is the i-th wall thickness.
Using a nearest neighbor (NN) search, the algorithm identifies walls within a distance threshold ( t d ) . For each wall, nearby candidates are found by comparing start and end points in 3D space. Intersections between two walls (Equation (9)) are obtained by solving the parametric line equations and equating r 1 t = r 2 s   and solving for the parameters ( t ) and ( s ) . If the parameters are within the valid range ( 0 t , s 1 ), the intersection lies within the bounds of both wall segments.
r 1 t = p start , 1 + t p end , 1 p start , 1 r 2 s = p start , 2 + s p end , 2 p start , 2
To refine wall relationships, orthogonal intersections are computed at endpoints using offsets   p start , p end   to detect potential overlaps or extensions. Distances from candidate intersections to wall axes are then evaluated and those exceeding the threshold ( t int ) are excluded, ensuring the accuracy of the connections. This ensures accurate representation of intersections, alignments, and overlaps, which are critical for reconstructing a realistic structural model (Figure 5). While transforming models from the Scan-to-BIM process into IFC format is supported, it remains optional within the proposed pipeline.

3.3.3. Openings ( O ) Class: Windows and Doors

The openings reconstruction process begins with the topological reconstruction of walls. From each wall’s centroid and main axis, orthographic images I x , y of the O class point cloud are generated via ray casting 1 m on both sides. These 2D projections capture wall surface details for object detection, with missing pixels filled black for consistency.
Object detection outputs bounding boxes x min ,   y min ,   x max ,   y max , where the corners define the detected region. Width w = x max x min and w = x max x min are calculated, incorporating a framing threshold t h f   =   0.07   m . Bounding boxes are overlaid on the orthographic images to annotate doors and windows, then mapped back into 3D space using known camera positions. Openings are reconstructed by sampling between start x 1 , y 1 , z 1 and end x 2 , y 2 , z 2 points and aligned with the wall’s axis. To ensure geometric robustness in downstream operations, such as surface Boolean intersections and wall embedding, each opening is slightly thickened along the normal direction.
Finally, each opening area is computed and stored for further thermal-envelope analyses. Openings are detected in 2D with GroundingDINO (39) (Section 3.2), backprojected using known camera poses, and intersected with the wall plane to obtain the 3D aperture polygon. The operable area is then obtained by insetting this polygon by a uniform frame thickness t f r = 0.075   m to derive the clear glazed portion within the fixed frame.

3.4. Scan-to-BEM

To support the detection of devices and energy-related systems, a lightweight interactive tool (Figure 6) was developed featuring a 3D point cloud viewport, with navigation and class selection for lighting, radiators, HVAC, and sensors. Device instances are identified by user clicks within the scene, followed by a ray casting and nearest-neighbor search in a k-d tree to snap to the closest in-class point within a radius of r snap = 0.10   m . Each confirmed instance is stored as an object containing the class name, element ID, and global coordinates x , y , z in meters.
Using the captured coordinates, each device is automatically assigned to the corresponding room or thermal zone through a point-in-solid spatial query; in cases of multiple overlaps, the assignment is made to the room with the largest volumetric intersection. A minimum inter-device distance per class (e.g., d m i n = 0.25   m for luminaires) is enforced to prevent duplicate selections.
The final annotation can be exported as a JSON file containing class_name, element_id, and coordinates, with optional room_id and confidence fields. Additional attributes related to energy performance, such as nominal power for radiators or seasonal efficiency for HVAC units, can be integrated post export using external manufacturer libraries or Information Loading Dictionaries (ILDs).
This spatially aware annotation approach enables the precise localization and classification of energy-critical components that are often partially concealed within architectural elements, for example in recessed luminaires and embedded radiators, thereby improving the completeness and accuracy of energy performance modeling.

3.5. Topologic Map and Topologic Model

To enable accurate energy simulation and performance analysis, the workflow requires the generation of B-Rep models that comply with strict geometric and topological constraints, including [99] the following:
  • Watertightness: All surfaces enclosing interior volumes must form a completely closed shell, without gaps or discontinuities.
  • Consistency and Manifoldness: Shared surfaces between adjacent spaces must be consistently defined, ensuring topological correctness.
The Topologic B-Rep model, grounded in the Topologic Map [95] framework and extended into three dimensions, is specifically engineered to address current limitations by providing a structured workflow for generating watertight volumetric representations of indoor spaces, which is an essential prerequisite for accurate and reliable energy simulations. The workflow initiates with the geometric abstraction of walls using minimum Oriented Bounding Boxes (mOBBs), forming the foundational elements for topological reasoning and space reconstruction.
Let W = { w 1 , w 2 , , w n } denote the set of all wall instances. Each wall w i is described as a collection of 3D points (Equation (9)) derived from the classified point cloud data, as detailed in the preceding section, while for each wall w i , its mOBB is defined as a triplet as shown in Equation (11):
w i = { p 1 i , p 2 i , , p m i i } , p j i R 3
mOBB w i = R i , c i , e i
where
  • R i S O 3 is a 3D rotation matrix is the rotation matrix of the local frame;
  • c i   R 3 is the center of the bounding box;
  • e i = e x , i , e y , i , e z , i R 3 represents the extents of the bounding box along each local axis.
From each mOBB, the wall centerline is obtained using the midpoints of its shorter edges as endpoints. Principal Component Analysis (PCA) on the wall points yields the dominant axis and the normal vector defines the local wall reference frame.
However, due to imperfections in the classification phase, occlusions in the surveying phase or the presence of columns at the intersection’s points, adjacent walls may not intersect and explicit joint points can be missing. A loop-based refinement computes candidate intersections, filters them using proximity thresholds and wall bounding box masks, and then consolidates vertices by snapping endpoints within a spatial tolerance, ensuring geometric consistency and topological correctness (Figure 7).
The process begins by generating wall buffers ( B w i ; δ )   for each wall instance w i   in W , where each buffer is defined as Equation (12):
B w i ;   δ = x R 2 dist x , centerline w i δ
where δ represents the buffer radius. Candidate connection segments are then computed from pairwise buffer intersections (Equation (13)):
S i , j = B w i ;   δ B w j ;   δ ,   for   all     i   j .
These segments are filtered by evaluating their angular similarity, where for each pair of direction vectors d i and d j , the angle is defined as Equations (14a) and (14b) [100]:
θ i , j = arccos d i d j | d i | | d j | , θ i , j 0 , π .
θ i , j =   d i   ,   d j
A candidate segment is retained only if θ i , j < ϵ θ , where ϵ θ   is the angular tolerance (here ϵ θ   =   5 ° ). The resulting intersection points v k R 2 are then refined by merging any pair v a , v b   such that | v a v b | < ϵ d , to avoid topological noise due to overly dense point distributions. Then, from the filtered and refined connections, a Doubly Connected Edge List (DCEL) is constructed by representing each segment s k = v a , v b   as two half-edges. In Equation (15), the arrow notation e . g . , v a v b denotes a directed half-edge from vertex v a to vertex v b in the topological graph. Each undirected edge is represented by two opposite half-edges
h e k = v a v b and   h e k = v b v a ,
which encode adjacency and orientation between connected surfaces.
These half-edges are linked by defining h e k ,   n e x t . next = h e k + 1   and h e k + 1 . prev = h e k , forming closed loops used to generate topological faces (Equation (16)):
f m = { h e m 1 , h e m 2 , , h e m p } ,   where   h e m p . next = h e m 1 .
Perimeter walls may sometimes be discontinuous and some segments may appear open or incomplete (Figure 7a), due to the absence of closing walls or missing boundary openings. To restore enclosure, closure surfaces are introduced: auxiliary zero thickness faces inserted only to seal perimeter gaps and produce a watertight envelope for analysis; they are synthetic and excluded from structural computations. A radial ray casting algorithm then constructs the outer boundary (Figure 8):
(i)
Emit rays uniformly from the building centroid over 360° at a fixed angular step; record each ray’s farthest valid wall hit and wall ID.
(ii)
Use filter hits to remove outliers/duplicates and retain points consistent with the exterior boundaries.
(iii)
Sort the remaining points counterclockwise by polar angle about the centroid.
(iv)
Detect gaps (distance/disconnect thresholds) and insert bridging segments via a Convex-Hull–based completion to obtain a continuous boundary polygon.
This robust structure enables the definition of complex room geometries, including both convex and concave configurations, based on the previously extracted connection segments. Finally, the height boundaries of the volumetric spaces are defined based on the floor and ceiling nodes, specifically using the previously computed values z m i n   and z m a x (Equation (2)), which represent the vertical extent of the enclosed room geometry.
This topological model is finally restituted in 3D, providing a coherent and watertight geometric structure that can be directly used for spatial reasoning and energy-related simulations (Figure 9).

3.6. The Multilevel Geometric Digital Twin (gDT)

The multilevel framework combines (i) volumetric (Topologic Solid model, TSM) and (ii) surface-based (B-Rep model) representations to capture the complexity of indoor environments. Both models originate from the 3D transformation of the 2D Topologic Map [101], embedding metadata within the solid nodes. In the TSM, a DCEL-based structure generates watertight 2D room volumes from segment connectivity and refined intersection logic, which are then extruded along the z axis with base and top boundaries set by the level node elevations (Equation (3)).
The (ii) B-Rep model semantically labels wall, ceiling, and floor surfaces using the Topologic Map 2D segments, while openings detected in the classification are matched to their host wall w i via geometric and orientation-aware projection. Each opening O j is mapped to a local 2D wall frame derived from the wall face’s principal axes and rotation matrices, filtered by area (thArea > 0.10 m2) and subtracted by Boolean difference to yield walls with precise holes (Equation (17)).
W hls = W i j = 1 n O j
The result is a hierarchical set of oriented polygons representing (i) plain wall surfaces, (ii) floors and ceilings, (iii) wall surfaces with openings, and (iv) individual opening geometries. Except for minor watertightness adjustments during room creation (e.g., closure walls), components retain a consistent, traceable identity (element_id) from point cloud detection through the final models.
This integrated representation forms the geometric Digital Twin (gDT), a layered, semantically enriched 3D model that unifies the following:
(i)
Volumetric Model (TSM): Watertight, room-level spatial cells (CellComplex model) for thermal zoning and energy simulation.
(ii)
Thematic Surface Model (B-Rep): Walls, floors, ceilings, and openings, each semantically annotated with material, function, and energy performance data.
(iii)
Energy System Elements and Devices: Geometric-topological representations and spatial registration of sensors and devices, enabling semantic graph-based reasoning and zone-level device mapping.
(iv)
Ancillary/Building physics data: A graph structure embedding construction layers, material properties, and performance attributes.
Finally, thanks to its multi-scale graph-based structure (Figure 10), the gDT supports semantic queries, interoperability, energy simulation and real-time deployment for smart building design, monitoring, and optimization. During faults, it pinpoints affected devices for faster diagnostics and maintenance.

4. Results

The following sections present results from the geometric gDT framework, demonstrating spatial semantic integration, device to space mapping, and energy-oriented reasoning. To validate the method, heterogeneous datasets from Terrestrial Laser Scanning and Mobile Laser Scanning were processed, spanning residential (House 01), educational (Block N, Gent, Belgium; PEA School, Italy), and heritage contexts (Santa Chiara, Trento, Italy). The datasets have comparable point cloud densities but differ in acquisition features and sensor configurations. A summary of these characteristics, along with information on whether indoor sensor data are available, is reported in Table 1, providing diverse conditions for evaluating the reconstruction pipeline.
Given the multi-stage nature of the Scan-to-EDTs framework, each step undergoes a consistent geometric assessment using oriented-bounding-box (OBB) matching with an IoU (Equation (18)) threshold (τ) and reporting Precision, Recall and F1 (Equation (19)), along with mean per-class accuracy (mAcc, Equation (20)) and mean per-class Intersection-over-Union (mIoU, Equation (21)).
IoU = A B A B
where
  • A   and   B are the predicted and ground truth (GT) oriented bounding boxes (OBBs);
  • A B is the intersection volume;
  • A B is the union volume.
F 1 = 2 T P 2 T P + F P + F N
where
  • TP are true positives, predicted OBBs correctly matched to a GT OBB;
  • FP are false positives, predicted OBBs not matched to any GT OBB;
  • FN are false negatives, GT OBBs not matched by any prediction.
Acc c = T P c T P c + F N c , mAcc = 1 C c = 1 C Ac c c
where c indexes classes, C is the number of classes; T P c and F N c are true and false negatives for class c (with positives/negatives defined per class).
Io U c = T P c T P c + F P c + F N c , mIoU = 1 C c = 1 C Io U c
Predictions are matched one-to-one to ground truth (GT) OBBs using IoU thresholds τ 0.05 ,   0.10 ,   0.15 . A prediction with I o U τ to an unmatched GT is counted as a TP; unmatched predictions are FP and unmatched GT instances are FN.

4.1. Data Classification

Semantic classification is not the primary focus of this research; nevertheless, the segmentation results are consistent with the ones reported on [94]. Point cloud segmentation uses Point Transformer v3 [26] with an internally pretrained model [94], taking point coordinates, RGB when available, and surface normals as inputs. In parallel, 2D image-based object detection follows the method in Section 3.1 and Section 3.2 and, mostly, in [94].
The two outputs are fused to ensure consistency within the 3D reconstruction workflow and results are reported in Table 2. The table indicates strong performance for major structural elements, such as floors, ceilings, and walls, that reach mIoU > 85% and mAcc > 89%, with low variance, reflecting limited variability and good representation in training data. In contrast, columns, doors, and windows score lower (mIoU = 44.1%, 51.6%, 58.8%) due to smaller size, occlusions, and class imbalance. Late fusion of point- and image-based outputs improves consistency for windows, confirming the value of multimodal integration. Overall, accuracy is sufficient for reliable topology reconstruction and energy-zone definition in subsequent pipeline stages.

4.2. Scan-to-BIM Process: Reconstruction Results

The following section presents results of the Scan-to-BIM workflow for the solid model (Figure 4, step A). A per-class reconstruction strategy (Section 3.3) [94] reconstructs each semantic class separately, allowing class-specific geometric assumptions and constraints, such as axis alignment for walls and columns, which leads to higher precision and completeness. Performance is evaluated using mean Intersection over Union (mIoU) and F1 score, defined in Equations (17) and (18), based on the spatial overlap between predicted elements and minimum Oriented Bounding Boxes (mOBBs) extracted from a manually segmented and classified point cloud that serves as instance level semantic ground truth.
Per class metrics (walls, columns, doors, etc.) are summarized in Table 3 and visualized in Figure 11. The overall mesh-to-cloud assessment is reported in Figure 12, last column, where the color bar represents the visual error distribution (ranging from −30 cm to +30 cm). Walls, benefiting from regular geometry and strong structural features, achieve consistently high accuracy (mIoU: 88.6–96.3%; F1: 85.2–94.0%). In contrast, doors and windows, being morphologically similar, frequently occluded, and often affected by clutter, yield moderate accuracy (mIoU: 67.8–77.3%; F1: 69.4–78.2%).
Notably, columns, which are generally a structurally sparse and spatially narrow class, show precision (mIoU: 75.1–78.3%; F1: 89.0–91.2%), though they are absent in certain datasets (House 01 or Block N). Performance varies with structural diversity, noise, occlusions, and class balance, with School PEA and Santa Chiara demonstrating the most consistency. A higher F1 than mIoU indicates reliable detection with some spatial imprecision, such as OBB misalignment. For House 01, separate reconstructions of the ground and first floor plans are shown. Currently, the method supports reconstructing one floor plan at a time. However, by processing each floor individually and subsequently merging them, a complete multi-level reconstruction can still be achieved, as demonstrated for House 01 and Block N.
Overall, the average distances remain low (0.02–0.12 m), with modest standard deviations, indicating isolated rather than systemic errors, and the accuracy is adequate for energy modeling that depends mainly on walls, floors, and ceilings.
Regarding energy-related elements, detection is performed manually via the tool described in Section 3.4. As it relies on visual interpretation rather than automated inference, no quantitative metrics (e.g., Accuracy or Recall) are provided.

4.3. Topologic Solid (TSM) and B-Rep Model Results

The Topologic Solid model (TSM) and B-Rep results are shown in Figure 13. Evaluation results (Table 4, Figure 14) show that TSMs (step B), also referred to as the CellComplex model, consistently outperform B-Rep models (step C) across all metrics. The accuracy of the surface objects of the B-Rep models was assessed against manually generated GT models, with a particular focus on the mean Intersection over Union (mIoU) parameter. Figure 14 illustrates the evaluation bounding boxes applied for the B-Rep model comparison, using a 0.065 m threshold, specifically for the evaluation of walls only.
Both models achieve high precision (TSM: 0.93–0.98, B-Rep: 0.74–0.95), confirming geometric accuracy when overlaps occur. However, recall is lower for B-Rep models (0.57–0.90) compared to TSMs (0.87–0.96), leading to reduced F1 scores (B-Rep: 0.65–0.89, TSM: 0.90–0.97) and IoU values (B-Rep: 0.72–0.95, TSM: 0.93–0.98). These discrepancies stem from differences in modeling strategy: TSMs represent enclosed volumes, while B-Rep models are surface-based and derived from wall centerlines.
Since GT models define walls by outer surfaces, consistent with energy modeling conventions, B-Rep models, though watertight and structurally correct via DCEL, yield lower recall, F1, and IoU results under voxel-based evaluation.
This effect is most evident in complex datasets. The PEA School, with dense layouts of small rooms and fine-grained partitions, produces the lowest B-Rep scores despite structural correctness.
Likewise, the Santa Chiara dataset, with irregular historic architecture and non-orthogonal geometries, presents challenges for B-Rep modeling and DCEL completion. To account for spatial discrepancies between centerline-based wall B-Rep modeling and the ground truth (GT) external wall surfaces defined by modeling standards, a buffer threshold (th) is applied (Figure 15).
This prevents accurate reconstructions from being unfairly penalized during evaluation, especially in IoU (Equation (18)), F1 (Equation (19)), Precision (Equation (22)), and Recall (Equation (23)). In contrast, TSM defines enclosed volumes explicitly, aligning more naturally with the GT evaluation models and achieving consistently higher scores.
Precision = p P g G , | p g | < τ P
Recall = g G p P , | g p | < τ G
where
  • P is the set of predicted points p P ;
  • G is the set of ground truth points g G ;
  • | p g | is the Euclidean distance between p P and g G ;
  • τ is the distance threshold (th).
Despite these discrepancies, the B-Rep results are adequate for energy performance applications because they preserve dimensions, structural logic, semantic clarity, and topological integrity, all of which are essential for simulation and analysis. Furthermore, the B-Rep models have been visually verified in FME® [101] to confirm model consistency, correct surface orientation, and overall topological validity (Figure 15).
The workflow yields a multi-level DT in which the TSM and B-Rep model coexist and are linked to the graph node structure. The DT is exportable to gbXML, enabling interoperable exchange with building performance simulation environments (Figure 16).

4.4. Energy Simulations

To demonstrate the feasibility of producing a simulation-ready DTL2 (Section 2.8), a preliminary simulation was performed using the model generated by the Scan-to-EDTs workflow in EnergyPlus [14]. The simulation employs an epJSON input file generated from the model output, real-world survey data, and standard simulation parameters.
This file was created using the first version of our custom-built parser–transformer–writer algorithm (Figure 17), developed in-house to process the tailored output.
This proof of concept validates the end-to-end integration of the pipeline, as the successful execution of energy simulations from the generated outputs confirms that the reconstructed zones are geometrically watertight, and the exported epJSON models are structurally consistent with all required inputs, while further robustness improvements are still needed for large-scale deployment. The pipeline also outputs models compatible with the latest CityGML Energy ADE 3.0 (GitHub: https://github.com/tudelft3d/Energy_ADE, URL accessed on 20 August 2025) [102], showing that the semantically enriched topological model maps cleanly to other formats and can upgrade existing single-zone CityGML urban models with multi-zone BEMs toward a more fine-grained UBEM. By embedding volumetric, thematic surface and geometric representations for visualization, the approach enables DTL2 and open platforms to support real-time or historical data display. Together, these results indicate that the end-to-end Scan-to-EDTs pipeline is operational. The interface (Figure 18) supports detailed inspection of the building model, including spatial hierarchies, thermal properties, material layers, and construction details, and it enables testing of retrofit scenarios by varying U values and integrating real-time monitoring data when available.

5. Conclusions and Future Research

This research presents a scalable 3D framework that produces simulation-ready Energy Digital Twins (EDTs) directly from unstructured point clouds, eliminating intermediate manual or BIM stages. The hierarchical schema encodes room volumes, semantic surfaces, and geometric elements in a unified graph structure, ensuring watertight, topologically valid, and semantically coherent models. The resulting EDTs support energy simulation, semantic querying, graph-based navigation, and real-time device integration, with seamless export to B-Rep-based gbXML and CityGML (with Energy ADE) and, optionally, CSG-based IFC for interoperability.
Several innovative aspects distinguish this research:
  • Deriving simulation-ready Energy Digital Twins (EDTs) directly from raw point clouds, using a graph-based native architecture that converts unstructured surveys into concrete, topologically consistent models without manual stages or proprietary tools.
  • Delivering watertight, simulation ready models with consistent topology and semantics, including adjacency, containment, connectivity, and energy attributes, enabling reliable reasoning, efficient querying, and seamless interoperability.
  • Unifying solid and B-Rep representations in a single multi-level gDT and integrated graph, enabling simulations, monitoring, IoT linking, optimization, and cross-scale navigation with consistent behavior across levels.
  • Enabling standards-based exchange via gbXML, CityGML and, optionally, IFC, ensuring standards-compliant interoperability while avoiding redundant steps and minimizing data loss across energy and BIM workflows.
In the literature, this paper links (1) point cloud semantics, (2) graph-based building models, and (3) BEM interoperability (via gbXML/IFC); unlike pipelines that rely on manual BIM authoring or mesh cleanup, it directly maps survey data into an energy-aware graph that preserves adjacency, containment, and connectivity and, by supporting both solid and B-Rep in a single multi-level structure, complements Scan-to-BIM while adding a direct-to-BEM route for simulation.
While the pipeline is effective end-to-end, some components can be strengthened. The classification and instance-segmentation modules occasionally introduce label noise, which can affect the stability of the overall workflow. Despite its practical efficiency, the Scan-to-BEM pipeline still relies on interactive elements, including manual device placement, which may introduce user-dependent variability into the results. Certain geometric simplifications (e.g., deriving exterior walls from centerlines) can lead to small deviations, especially at complex façades. Finally, because this is a point-cloud-based workflow, performance remains sensitive to input coverage and quality (e.g., occlusions or sparse scans).
Future research will focus on the geometric reconstruction pipeline for energy-model compliance: (i) improving classification and instance segmentation using state-of-the-art networks; (ii) placing external walls on the true exterior face (replacing the centerline placement used in the presented workflow); (iii) optimizing the epJSON parser–transformer–writer algorithm for parameter customization; (iv) refining the linked Informational Load Dictionaries (ILDs) with explicit schema definitions and improved interoperability mappings; (v) implement better interoperability with the evolving CityGML Energy ADE [102] to inject multi-zone BEMs into single-zone city models and to export gbXML for seamless use in BPS tools; (vi) add high-fidelity IFC mapping to capture rich BIM semantics; and finally, (vii) full automation for simulation-ready BEM generation will streamline deployment across diverse energy-analysis platforms.

Author Contributions

Conceptualization: O.R., F.R.; Methodology: O.R., M.B., G.A., K.A.O. and F.R.; Investigation: M.B., G.A. and K.A.O.; Resources: M.B., G.A. and K.A.O.; Software: O.R.; Technical Implementation: O.R., M.B., G.A., K.A.O. and F.R.; Data Curation: all authors; Writing—Original Draft: O.R.; Writing—Review and Editing: all authors; Formal Analysis: all authors; Visualization: O.R.; Supervision: M.B., G.A., K.A.O., E.M.F. and F.R.; Project Administration: E.M.F. and F.R. All authors have read and agreed to the published version of the manuscript.

Funding

The research was carried out within the PhD program in Industrial Innovation at the University of Trento (Italy)—38th cycle, with the support of a scholarship co-financed by Edilvi and the Italian Ministerial Decree no. 352 of 9th April 2022, based on the NRRP, CUP E66E22000050008 —funded by the European Union—NextGenerationEU—Mission 4 “Education and Research”, Component 2 “From Research to Business”, Investment 3.3.

Data Availability Statement

Raw data are available upon request to the author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Available online: https://www.iea.org/energy-system/buildings (accessed on 12 August 2025).
  2. Gouveia, J.P.; Aelenei, L.; Aelenei, D.; Ourives, R.; Bessa, S. Improving the Energy Performance of Public Buildings in the Mediterranean Climate via a Decision Support Tool. Energies 2024, 17, 1105. [Google Scholar] [CrossRef]
  3. Tsemekidi Tzeiranaki, S.; Bertoldi, P.; Dagostino, D.; Castellazzi, L.; Maduta, C. Energy Consumption and Energy Efficiency Trends in the EU, 2000–2022; Publications Office of the European Union: Luxembourg, 2025. [Google Scholar]
  4. European Parliament. Fact Sheets on the European Union: Renewable Energy. European Parliament. 2024. Available online: https://www.europarl.europa.eu/factsheets/en/sheet/70/renewable-energy (accessed on 12 August 2025).
  5. Mendes, V.F.; Cruz, A.S.; Gomes, A.P.; Mendes, J.C. A Systematic Review of Methods for Evaluating the Thermal Performance of Buildings through Energy Simulations. Renew. Sustain. Energy Rev. 2024, 189, 113875. [Google Scholar] [CrossRef]
  6. Liebenberg, M.; Jarke, M. Information Systems Engineering with Digital Shadows: Concept and Use Cases in the Internet of Production. Inf. Syst. 2023, 114, 102182. [Google Scholar] [CrossRef]
  7. Hu, X.; Olgun, G.; Assaad, R.H. An Intelligent BIM-Enabled Digital Twin Framework for Real-Time Structural Health Monitoring Using Wireless IoT Sensing, Digital Signal Processing, and Structural Analysis. Expert Syst. Appl. 2024, 252, 124204. [Google Scholar] [CrossRef]
  8. Kreuzer, T.; Papapetrou, P.; Zdravkovic, J. Artificial Intelligence in Digital Twins—A Systematic Literature Review. Data Knowl. Eng. 2024, 151, 102304. [Google Scholar] [CrossRef]
  9. Armijo, A.; Zamora-Sánchez, D. Integration of Railway Bridge Structural Health Monitoring into the Internet of Things with a Digital Twin: A Case Study. Sensors 2024, 24, 2115. [Google Scholar] [CrossRef]
  10. Cummins, L.; Sommers, A.; Ramezani, S.B.; Mittal, S.; Jabour, J.; Seale, M.; Rahimi, S. Explainable Predictive Maintenance: A Survey of Current Methods, Challenges and Opportunities. IEEE Access 2024, 12, 57574–57602. [Google Scholar] [CrossRef]
  11. Attaran, S.; Attaran, M.; Celik, B.G. Digital Twins and Industrial Internet of Things: Uncovering Operational Intelligence in Industry 4.0. Decis. Anal. J. 2024, 10, 100398. [Google Scholar] [CrossRef]
  12. Meddaoui, A.; Hain, M.; Hachmoud, A. The benefits of predictive maintenance in manufacturing excellence: A case study to establish reliable methods for predicting failures. Int. J. Adv. Manuf. Technol. 2023, 128, 3685–3690. [Google Scholar] [CrossRef]
  13. Asare, K.A.B.; Liu, R.; Anumba, C.J.; Issa, R.R.A. Real-World Prototyping and Evaluation of Digital Twins for Predictive Facility Maintenance. J. Build. Eng. 2024, 97, 110890. [Google Scholar] [CrossRef]
  14. Liu, Z.; Li, M.; Ji, W. Development and Application of a Digital Twin Model for Net Zero Energy Building Operation and Maintenance Utilizing BIM-IoT Integration. Energy Build. 2025, 328, 115170. [Google Scholar] [CrossRef]
  15. Naeem, A.; Ho, C.O.; Kolderup, E.; Jain, R.K.; Benson, S.; de Chalendar, J. EnergyPlus as a computational engine for commercial building operational digital twins. Energy Build. 2025, 329, 115257. [Google Scholar] [CrossRef]
  16. Pavirani, F.; Gokhale, G.; Claessens, B.; Develder, C. Demand Response for Residential Building Heating: Effective Monte Carlo Tree Search Control Based on Physics-Informed Neural Networks. Energy Build. 2024, 311, 114161. [Google Scholar] [CrossRef]
  17. Mohammadi, S.; Aibinu, A.A.; Oraee, M. Legal and Contractual Risks and Challenges for BIM. J. Leg. Aff. Dispute Resolut. Eng. Constr. 2024, 16, 1. [Google Scholar] [CrossRef]
  18. Palha, R.P.; Hüttl, R.M.C.; da Costa e Silva, A.J. BIM Interoperability for Small Residential Construction Integrating Warranty and Maintenance Management. Autom. Constr. 2024, 166, 105639. [Google Scholar] [CrossRef]
  19. Yang, Y.; Pan, Y.; Zeng, F.; Lin, Z.; Li, C. A gbXML Reconstruction Workflow and Tool Development to Improve the Geometric Interoperability between BIM and BEM. Buildings 2022, 12, 221. [Google Scholar] [CrossRef]
  20. Yang, Z.; Tang, C.; Zhang, T.; Zhang, Z.; Doan, D.T. Digital Twins in Construction: Architecture, Applications, Trends and Challenges. Buildings 2024, 14, 2616. [Google Scholar] [CrossRef]
  21. Szpilko, D.; Fernando, X.; Nica, E.; Budna, K.; Rzepka, A.; Lăzăroiu, G. Energy in Smart Cities: Technological Trends and Prospects. Energies 2024, 17, 6439. [Google Scholar] [CrossRef]
  22. Mousavi, Y.; Gharineiat, Z.; Karimi, A.A.; McDougall, K.; Rossi, A.; Gonizzi Barsanti, S. Digital Twin Technology in Built Environment: A Review of Applications, Capabilities and Challenges. Smart Cities 2024, 7, 2595–2621. [Google Scholar] [CrossRef]
  23. Wang, W.; Xu, K.; Song, S.; Bao, Y.; Xiang, C. From BIM to Digital Twin in BIPV: A Review of Current Knowledge. Sustain. Energy Technol. Assess. 2024, 67, 103855. [Google Scholar] [CrossRef]
  24. Yavan, F.; Maalek, R.; Toğan, V. Structural Optimization of Trusses in Building Information Modeling (BIM) Projects Using Visual Programming, Evolutionary Algorithms, and Life Cycle Assessment (LCA) Tools. Buildings 2024, 14, 1532. [Google Scholar] [CrossRef]
  25. Wang, Q.; Chen, X.; Zhang, Z.; Meng, Y.; Shen, T.; Gu, Y. Masking Graph Cross-Convolution Network for Multispectral Point Cloud Classification. IEEE Trans. Geosci. Remote Sens. 2025, 63, 1–15. [Google Scholar] [CrossRef]
  26. Zhao, H.; Jiang, L.; Jia, J.; Torr, P.; Koltun, V. Point Transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 16259–16268. [Google Scholar]
  27. Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  28. Han, T.; Chen, Y.; Ma, J.; Liu, X.; Zhang, W.; Zhang, X.; Wang, H. Point Cloud Semantic Segmentation with Adaptive Spatial Structure Graph Transformer. Int. J. Appl. Earth Obs. Geoinf. 2024, 133, 104105. [Google Scholar] [CrossRef]
  29. Zhou, W.; Wang, Q.; Jin, W.; Shi, X.; He, Y. Graph Transformer for 3D Point Clouds Classification and Semantic Segmentation. Comput. Graph. 2024, 124, 104050. [Google Scholar] [CrossRef]
  30. Zhou, W.; Zhao, Y.; Xiao, Y.; Min, X.; Yi, J. TNPC: Transformer-Based Network for Point Cloud Classification. Expert Syst. Appl. 2024, 239, 122438. [Google Scholar] [CrossRef]
  31. Wang, L.; Huang, M.; Yang, Z.; Wu, R.; Qiu, D.; Xiao, X.; Li, D.; Chen, C. LBNP: Learning Features Between Neighboring Points for Point Cloud Classification. PLoS ONE 2025, 20, e0314086. [Google Scholar] [CrossRef] [PubMed]
  32. Xie, Y.; Tu, Z.; Yang, T.; Zhang, Y.; Zhou, X. EdgeFormer: Local Patch-Based Edge Detection Transformer on Point Clouds. Pattern Anal. Appl. 2025, 28, 11. [Google Scholar] [CrossRef]
  33. Thomas, H.; Tsai, Y.-H.; Barfoot, T.D.; Zhang, J. KPConvX: Modernizing Kernel Point Convolution with Kernel Attention. In Proceedings of the 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 16–22 June 2024; pp. 5525–5535. [Google Scholar]
  34. Wang, L.; Deng, H.; Li, S.; Liu, C. PPT: A Point Patch Transformer for Point Cloud Classification. Optica Open 2025, preprint. [Google Scholar] [PubMed]
  35. Park, Y.; Tran, D.D.T.; Kim, M.; Kim, H.; Lee, Y. SP2Mask4D: Efficient 4D Panoptic Segmentation Using Superpoint Transformers. In Proceedings of the 2025 International Conference on Electronics, Information, and Communication (ICEIC), Osaka, Japan, 19–22 January 2025; pp. 1–4. [Google Scholar]
  36. Fan, Y.; Wang, Y.; Zhu, P.; Hui, L.; Xie, J.; Hu, Q. Uncertainty-Aware Superpoint Graph Transformer for Weakly Supervised 3D Semantic Segmentation. IEEE Trans. Fuzzy Syst. 2025, 33, 1899–1912. [Google Scholar] [CrossRef]
  37. Mei, G.; Riz, L.; Wang, Y.; Poiesi, F. Vocabulary-Free 3D Instance Segmentation with Vision and Language Assistant. arXiv 2024, arXiv:2408.10652. [Google Scholar] [CrossRef]
  38. Kirillov, A.; Mintun, E.; Ravi, N.; Mao, H.; Rolland, C.; Gustafson, L.; Xiao, T.; Whitehead, S.; Berg, A.C.; Lo, W.; et al. Segment Anything. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 1–6 October 2023; pp. 3992–4003. [Google Scholar]
  39. Liu, S.; Zeng, Z.; Ren, T.; Li, F.; Zhang, H.; Yang, J.; Li, C.; Yang, J.; Su, H.; Zhu, J.; et al. Grounding DINO: Marrying DINO with Grounded Pre-Training for Open-Set Object Detection. arXiv 2023, arXiv:2303.05499. [Google Scholar]
  40. Zhang, W.; Pang, J.; Chen, K.; Loy, C.C. K-Net: Towards Unified Image Segmentation. arXiv 2021, arXiv:2106.14855. [Google Scholar] [CrossRef]
  41. Qin, Z.; Liu, J.; Zhang, X.; Tian, M.; Zhou, A.; Yi, S.; Li, H. Pyramid Fusion Transformer for Semantic Segmentation. arXiv 2022, arXiv:2201.04019. [Google Scholar] [CrossRef]
  42. Cheng, B.; Schwing, A.G.; Kirillov, A. Per-Pixel Classification is Not All You Need for Semantic Segmentation. Adv. Neural Inf. Process. Syst. 2021, 34, 17864–17875. [Google Scholar]
  43. Wu, Z.; Gan, Y.; Xu, T.; Wang, F. Graph-Segmenter: Graph Transformer with Boundary-aware Attention for Semantic Segmentation. Front. Comput. Sci. 2023, 18, 185327. [Google Scholar] [CrossRef]
  44. Dai, K.; Jiang, Z.; Xie, T.; Wang, K.; Liu, D.; Fan, Z.; Li, R.; Zhao, L.; Omar, M. SOFW: A Synergistic Optimization Framework for Indoor 3D Object Detection. IEEE Trans. Multimed. 2025, 27, 637–651. [Google Scholar] [CrossRef]
  45. Lin, T.; Yu, Z.; McGinity, M.; Gumhold, S. An Immersive Labeling Method for Large Point Clouds. Comput. Graph. 2024, 124, 104101. [Google Scholar] [CrossRef]
  46. Fol, C.R.; Shi, N.; Overney, N.; Murtiyoso, A.; Remondino, F.; Griess, V.C. 3D Dataset Generation Using Virtual Reality for Forest Biodiversity. Int. J. Digit. Earth 2024, 17, 2422984. [Google Scholar] [CrossRef]
  47. Shi, X.; Wang, S.; Zhang, B.; Ding, X.; Qi, P.; Qu, H.; Li, N.; Wu, J.; Yang, H. Advances in Object Detection and Localization Techniques for Fruit Harvesting Robots. Agronomy 2025, 15, 145. [Google Scholar] [CrossRef]
  48. ManyCore Research Team. SpatialLM: Large Language Model for Spatial Understanding. GitHub Repository. 2025. Available online: https://github.com/manycore-research/SpatialLM (accessed on 28 March 2025).
  49. Yang, Q.; Zhao, Y.; Cheng, H. MMLF: Multi-Modal Multi-Class Late Fusion for Object Detection with Uncertainty Estimation. arXiv 2024, arXiv:2410.08739. [Google Scholar]
  50. Ren, D.; Li, J.; Wu, Z.; Guo, J.; Wei, M.; Guo, Y. MFFNet: Multimodal Feature Fusion Network for Point Cloud Semantic Segmentation. Vis. Comput. 2024, 40, 5155–5167. [Google Scholar] [CrossRef]
  51. Agostinho, L.; Pereira, D.; Hiolle, A.; Pinto, A. TEFu-Net: A Time-Aware Late Fusion Architecture for Robust Multi-Modal Ego-Motion Estimation. Robot. Auton. Syst. 2024, 177, 104700. [Google Scholar] [CrossRef]
  52. Ji, A.; Chew, A.W.Z.; Xue, X.; Zhang, L. An Encoder-Decoder Deep Learning Method for Multi-Class Object Segmentation from 3D Tunnel Point Clouds. Autom. Constr. 2022, 137, 104187. [Google Scholar] [CrossRef]
  53. Samadzadegan, F.; Toosi, A.; Dadrass Javan, F. A Critical Review on Multi-Sensor and Multi-Platform Remote Sensing Data Fusion Approaches: Current Status and Prospects. Int. J. Remote Sens. 2024, 46, 1327–1402. [Google Scholar] [CrossRef]
  54. Schuegraf, P.; Schnell, J.; Henry, C.; Bittner, K. Building Section Instance Segmentation with Combined Classical and Deep Learning Methods. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2022, II, 407–414. [Google Scholar] [CrossRef]
  55. Liu, Y.; Huang, H.; Gao, G.; Ke, Z.; Li, S.; Gu, M. Dataset and Benchmark for As-Built BIM Reconstruction from Real-World Point Cloud. Autom. Constr. 2025, 173, 106096. [Google Scholar] [CrossRef]
  56. Bassier, M.; Vergauwen, M. Clustering of Wall Geometry from Unstructured Point Clouds Using Conditional Random Fields. Remote Sens. 2019, 11, 1586. [Google Scholar] [CrossRef]
  57. Skoury, L.; Leder, S.; Menges, A.; Wortmann, T. Digital Twin Architecture for the AEC Industry: A Case Study in Collective Robotic Construction. In Proceedings of the MODELS Companion ‘24: Proceedings of the ACM/IEEE 27th International Conference on Model Driven Engineering Languages and Systems; Linz, Austria, 22–27 September 2024, Volume HS 18.
  58. Vega Torres, M.A.; Braun, A.; Borrmann, A. BIM-SLAM: Integrating BIM Models in Multi-session SLAM for Lifelong Mapping using 3D LiDAR. In Proceedings of the International Symposium on Automation and Robotics in Construction ISARC, Chennai, India, 4–7 July 2023. [Google Scholar]
  59. Huang, H.; Qiao, Z.; Yu, Z.; Liu, C.; Shen, S.; Zhang, F.; Yin, H. SLABIM: A SLAM-BIM Coupled Dataset in HKUST Main Building. arXiv 2025, arXiv:2502.16856. [Google Scholar] [CrossRef]
  60. Lin, S.; Duan, L.; Jiang, B.; Liu, J.; Guo, H.; Zhao, J. Scan vs. BIM: Automated Geometry Detection and BIM Updating of Steel Framing through Laser Scanning. Autom. Constr. 2025, 170, 105931. [Google Scholar] [CrossRef]
  61. Bahreini, F.; Nasrollahi, M.; Taher, A.; Hammad, A. Ontology for BIM-Based Robotic Navigation and Inspection Tasks. Buildings 2024, 14, 2274. [Google Scholar] [CrossRef]
  62. Hu, X.; Assaad, R.H. A BIM-Enabled Digital Twin Framework for Real-Time Indoor Environment Monitoring and Visualization by Integrating Autonomous Robotics, LiDAR-Based 3D Mobile Mapping, IoT Sensing, and Indoor Positioning Technologies. J. Build. Eng. 2024, 86, 108901. [Google Scholar] [CrossRef]
  63. Yarovoi, A.; Cho, Y.K. Review of Simultaneous Localization and Mapping (SLAM) for Construction Robotics Applications. Autom. Constr. 2024, 162, 105344. [Google Scholar] [CrossRef]
  64. De Geyter, S.; Bassier, M.; De Winter, H.; Vergauwen, M. Review of Window and Door Type Detection Approaches. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2022, XLVIII-2/W1, 65–72. [Google Scholar] [CrossRef]
  65. van der Vaart, J.; Stoter, J.; Agugiaro, G.; Arroyo Ohori, K.; Hakim, A.; El Yamani, S. Enriching Lower LoD 3D City Models with Semantic Data Computed by the Voxelisation of BIM Sources. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2024, X-4/W5, 297–308. [Google Scholar] [CrossRef]
  66. Bassier, M.; Vermandere, J.; De Geyter, S.; De Winter, H. GEOMAPI: Processing Close-Range Sensing Data of Construction Scenes with Semantic Web Technologies. Autom. Constr. 2024, 164, 105454. [Google Scholar] [CrossRef]
  67. Tran, H.; Khoshelham, K. Procedural Reconstruction of 3D Indoor Models from Lidar Data Using Reversible Jump Markov Chain Monte Carlo. Remote Sens. 2020, 12, 838. [Google Scholar] [CrossRef]
  68. Walch, A.; Szabo, A.; Steinlechner, H.; Ortner, T.; Gröller, E.; Schmidt, J. BEMTrace: Visualization-Driven Approach for Deriving Building Energy Models from BIM. IEEE Trans. Vis. Comput. Graph. 2025, 31, 240–250. [Google Scholar] [CrossRef]
  69. Reddy, V.J.; Hariram, N.P.; Ghazali, M.F.; Kumarasamy, S. Pathway to Sustainability: An Overview of Renewable Energy Integration in Building Systems. Sustainability 2024, 16, 638. [Google Scholar] [CrossRef]
  70. Arowoiya, V.A.; Moehler, R.C.; Fang, Y. Digital Twin Technology for Thermal Comfort and Energy Efficiency in Buildings: A State-of-the-Art and Future Directions. Energy Built Environ. 2024, 5, 641–656. [Google Scholar] [CrossRef]
  71. Krispel, U.; Evers, H.; Tamke, M.; Ullrich, T. Data Completion in Building Information Management: Electrical Lines from Range Scans and Photographs. Vis. Eng. 2017, 5, 4. [Google Scholar] [CrossRef]
  72. Yeom, S.; Kim, J.; Kang, H.; Jung, S.; Hong, T. Digital Twin (DT) and Extended Reality (XR) for Building Energy Management. Energy Build. 2024, 323, 114746. [Google Scholar] [CrossRef]
  73. Han, F.; Du, F.; Jiao, S.; Zou, K. Predictive Analysis of a Building’s Power Consumption Based on Digital Twin Platforms. Energies 2024, 17, 3692. [Google Scholar] [CrossRef]
  74. Michalakopoulos, V.; Pelekis, S.; Kormpakis, G.; Karakolis, V.; Mouzakitis, S.; Askounis, D. Data-Driven Building Energy Efficiency Prediction Using Physics-Informed Neural Networks. In Proceedings of the 2024 IEEE Conference on Technologies for Sustainability (SusTech), Portland, OR, USA, 14–17 April 2024; pp. 84–91. [Google Scholar]
  75. Walczyk, G.; Ożadowicz, A. Building Information Modeling and Digital Twins for Functional and Technical Design of Smart Buildings with Distributed IoT Networks—Review and New Challenges Discussion. Future Internet 2024, 16, 225. [Google Scholar] [CrossRef]
  76. Kiavarz, H.; Jadidi, M.; Esmaili, P. A Graph-Based Explanatory Model for Room-Based Energy Efficiency Analysis Based on BIM Data. Front. Built Environ. 2023, 9, 1256921. [Google Scholar] [CrossRef]
  77. Wang, M.; Lilis, G.N.; Mavrokapnidis, D.; Katsigarakis, K.; Korolija, I.; Rovas, D. A Knowledge Graph-Based Framework to Automate the Generation of Building Energy Models Using Geometric Relation Checking and HVAC Topology Establishment. Energy Build. 2024, 325, 115035. [Google Scholar] [CrossRef]
  78. Available online: https://github.com/ladybug-tools (accessed on 16 April 2025).
  79. Zhou, Y.; Liu, J. Advances in Emerging Digital Technologies for Energy Efficiency and Energy Integration in Smart Cities. Energy Build. 2024, 315, 114289. [Google Scholar] [CrossRef]
  80. Liu, Z.; He, Y.; Demian, P.; Osmani, M. Immersive Technology and Building Information Modeling (BIM) for Sustainable Smart Cities. Buildings 2024, 14, 1765. [Google Scholar] [CrossRef]
  81. Jabi, W.; Soe, S.; Theobald, P.; Aish, R.; Lannon, S. Enhancing Parametric Design through Non-Manifold Topology. Des. Stud. 2017, 52, 96–114. [Google Scholar] [CrossRef]
  82. Yue, Y.; Kontogianni, T.; Schindler, K.; Engelmann, F. Connecting the Dots: Floorplan Reconstruction Using Two-Level Queries. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023. [Google Scholar]
  83. Aish, R.; Jabi, W.; Lannon, S.; Wardhana, N.; Chatzivasileiadi, A. Topologic: Tools to Explore Architectural Topology. In Proceedings of the Advances in Architectural Geometry 2018, Gothenburg, Sweden, 22–25 September 2018. [Google Scholar]
  84. Moradi, M.A.; Mohammadrashidi, O.; Niazkar, N.; Rahbar, M. Revealing Connectivity in Residential Architecture: An Algorithmic Approach to Extracting Adjacency Matrices from Floor Plans. Front. Archit. Res. 2024, 13, 370–386. [Google Scholar] [CrossRef]
  85. Han, J.; Lu, X.-Z.; Lin, J.-R. Unified Network-Based Representation of BIM Models for Embedding Semantic, Spatial, and Topological Data. arXiv 2025, arXiv:2505.22670. [Google Scholar]
  86. Allen, B.D. Digital twins and living models at NASA. In Proceedings of the Digital Twin Summit, Virtual, 3–4 November 2021. [Google Scholar]
  87. Nguyen, T.D.; Adhikari, S. The Role of BIM in Integrating Digital Twin in Building Construction: A Literature Review. Sustainability 2023, 15, 10462. [Google Scholar] [CrossRef]
  88. Tuhaise, V.V.; Tah, J.H.M.; Abanda, F.H. Technologies for Digital Twin Applications in Construction. Autom. Constr. 2023, 152, 104931. [Google Scholar] [CrossRef]
  89. Kang, T.W.; Mo, Y. A Comprehensive Digital Twin Framework for Building Environment Monitoring with Emphasis on Real-Time Data Connectivity and Predictability. Dev. Built Environ. 2024, 17, 100309. [Google Scholar] [CrossRef]
  90. Chen, X.; Pan, Y.; Gan, V.J.L.; Yan, K. 3D reconstruction of semantic-rich digital twins for ACMV monitoring and anomaly detection via scan-to-BIM and time-series data integration. Dev. Built Environ. 2024, 19, 100503. [Google Scholar] [CrossRef]
  91. Arsecularatne, B.P.; Rodrigo, N.; Chang, R. Review of reducing energy consumption and carbon emissions through digital twin in built environment. J. Build. Eng. 2024, 98, 111150. [Google Scholar] [CrossRef]
  92. Wysocki, O.; Schwab, B.; Biswanath, M.K.; Zhang, Q.; Zhu, J.; Froech, T.; Heeramaglore, M.; Hijazi, I.; Kanna, K.; Pechinger, M.; et al. TUM2TWIN: Introducing the Large-Scale Multimodal Urban Digital Twin Benchmark Dataset. arXiv 2025, arXiv:2505.07396. [Google Scholar]
  93. Faraji, A.; Arya, S.H.; Ghasemi, E.; Soleimani, H. Considering the Digital Twin as the Evolved Level of BIM in the Context of Construction 4.0. In Proceedings of the 1st National & International Conference on Architecture, Advanced Technologies and Construction Management (IAAC), Tehran, Iran, 15 November 2023. [Google Scholar]
  94. Roman, O.; Bassier, M.; De Geyter, S.; De Winter, H.; Farella, E.M.; Remondino, F. BIM Module for Deep Learning-Driven Parametric IFC Reconstruction. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2024, XLVIII-2/W8, 403–410. [Google Scholar] [CrossRef]
  95. Roman, O.; Mazzacca, G.; Farella, E.M.; Remondino, F.; Bassier, M.; Agugiaro, G. Towards Automated BIM and BEM Model Generation Using a B-Rep-Based Method with Topological Map. ISPRS Ann. Phot. Remote Sens. Spat. Inf. Sci. 2024, X-4, 287–294. [Google Scholar] [CrossRef]
  96. Wu, X.; Jiang, L.; Wang, P.-S.; Liu, Z.; Liu, X.; Qiao, Y.; Ouyang, W.; He, T.; Zhao, H. Point Transformer v3: Simpler Faster Stronger. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 17–21 June 2024; pp. 4840–4851. [Google Scholar]
  97. Varghese, R.; Sambath, M. YOLOv8: A Novel Object Detection Algorithm with Enhanced Performance and Robustness. In Proceedings of the International Conference on Advanced Data Engineering and Intelligent Computing Systems (ADICS), Chennai, India, 18–19 April 2024. [Google Scholar]
  98. Hahsler, M.; Piekenbrock, M. dbscan: Density-Based Spatial Clustering of Applications with Noise (DBSCAN) and Related Algorithms. J. Stat. Softw. 2023, 91, 1–30. [Google Scholar]
  99. Agugiaro, G.; Peters, R.; Stoter, J.; Dukai, B. Computing Volumes and Surface Areas Including Party Walls for the 3DBAG Data Set; TU Delft/3DGI: Delft, The Netherlands, 2023. [Google Scholar]
  100. Roman, O.; Farella, E.M.; Rigon, S.; Remondino, F.; Ricciuti, S.; Viesi, D. From 3D Surveying Data to BIM to BEM: The INCUBE Dataset. ISPRS Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2023, XLVIII-1/W3, 175–182. [Google Scholar] [CrossRef]
  101. FME Software. Copyright (c) Safe Software Inc. Available online: www.safe.com (accessed on 28 March 2025).
  102. Agugiaro, G.; Padsala, R. A proposal to update and enhance the CityGML Energy Application Domain Extension. ISPRS Ann. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2025, X-4/W6-2025, 1–8. [Google Scholar] [CrossRef]
Figure 1. Main workflow schema for bridging physical and digital assets in building DTs.
Figure 1. Main workflow schema for bridging physical and digital assets in building DTs.
Buildings 15 04060 g001
Figure 2. The traditional workflow to derive a DT passing from the BIM creation.
Figure 2. The traditional workflow to derive a DT passing from the BIM creation.
Buildings 15 04060 g002
Figure 3. The workflow towards gDT and DT: requisites and maturity levels for DTs.
Figure 3. The workflow towards gDT and DT: requisites and maturity levels for DTs.
Buildings 15 04060 g003
Figure 4. The proposed workflow for the Scan-to-EDTs process.
Figure 4. The proposed workflow for the Scan-to-EDTs process.
Buildings 15 04060 g004
Figure 5. (a) Small example of solid model reconstruction with corresponding point cloud, (b1) solid model and labeled point cloud, (b2) solid model visualization, (b3) detail of the building corner indentation, (b4) reconstructed IFC model with indentation detail.
Figure 5. (a) Small example of solid model reconstruction with corresponding point cloud, (b1) solid model and labeled point cloud, (b2) solid model visualization, (b3) detail of the building corner indentation, (b4) reconstructed IFC model with indentation detail.
Buildings 15 04060 g005
Figure 6. Results of the developed tool for manual energy-element detection and classification: (a) interface and lights, (b) lights, (c) radiators, (d) devices/sensors, (e) top view after detection.
Figure 6. Results of the developed tool for manual energy-element detection and classification: (a) interface and lights, (b) lights, (c) radiators, (d) devices/sensors, (e) top view after detection.
Buildings 15 04060 g006
Figure 7. Steps for intersections computation and edge refinement: (a) iteratively compute all the intersections; (b) retain only the overlapping segments; (c) clear and refine floorplan segments.
Figure 7. Steps for intersections computation and edge refinement: (a) iteratively compute all the intersections; (b) retain only the overlapping segments; (c) clear and refine floorplan segments.
Buildings 15 04060 g007
Figure 8. (a) Ray tracing for external segment identification; (b) external segment computation; (c) cleared edges and closure segments (in blue).
Figure 8. (a) Ray tracing for external segment identification; (b) external segment computation; (c) cleared edges and closure segments (in blue).
Buildings 15 04060 g008
Figure 9. Steps for floorplan definition: (a) Computed single edges, (b) labeled edges, and (c) room polygon for floorplan.
Figure 9. Steps for floorplan definition: (a) Computed single edges, (b) labeled edges, and (c) room polygon for floorplan.
Buildings 15 04060 g009
Figure 10. Enriched graphs through operations: (1) solid model with rooms (Topologic Solid model); (2) Topologic Solid model with energy-related elements; (3) Topologic B-Rep model with thematic surfaces; (4) graph of the full model with thematic surfaces and energy elements.
Figure 10. Enriched graphs through operations: (1) solid model with rooms (Topologic Solid model); (2) Topologic Solid model with energy-related elements; (3) Topologic B-Rep model with thematic surfaces; (4) graph of the full model with thematic surfaces and energy elements.
Buildings 15 04060 g010
Figure 11. Metrics results for solid-based elements reported for the per-class reconstruction.
Figure 11. Metrics results for solid-based elements reported for the per-class reconstruction.
Buildings 15 04060 g011
Figure 12. Scan-to-BIM reconstruction process: (a) Raw point cloud; (b) instance-segmented point cloud; (c) generated solid model output; (d) optional IFC wall-class reconstruction results; (e) evaluation metrics comparing the solid model to the input point cloud.
Figure 12. Scan-to-BIM reconstruction process: (a) Raw point cloud; (b) instance-segmented point cloud; (c) generated solid model output; (d) optional IFC wall-class reconstruction results; (e) evaluation metrics comparing the solid model to the input point cloud.
Buildings 15 04060 g012
Figure 13. All results obtained: column (a), topological, internal, external, and closure surfaces; column (b), solid model; column (c), complete topologic model; column (d), topologic model and devices; column (e), comprehensive graph of structural and energy-related elements in building.
Figure 13. All results obtained: column (a), topological, internal, external, and closure surfaces; column (b), solid model; column (c), complete topologic model; column (d), topologic model and devices; column (e), comprehensive graph of structural and energy-related elements in building.
Buildings 15 04060 g013
Figure 14. Graphic results for TSM and B-Rep model assessment across datasets.
Figure 14. Graphic results for TSM and B-Rep model assessment across datasets.
Buildings 15 04060 g014
Figure 15. Bottom left: FME checks of the topologic model (ceilings and floors, walls, openings, full model). Right: Wall bounding boxes for mIoU at th = 0.065 m.
Figure 15. Bottom left: FME checks of the topologic model (ceilings and floors, walls, openings, full model). Right: Wall bounding boxes for mIoU at th = 0.065 m.
Buildings 15 04060 g015
Figure 16. Multi-level gDT (ac) and its structure (b) are exported to gbXML (c) and revisualized as thermal zone solids (d1d3) and thematic surfaces (e1e3), demonstrating interoperability with building performance tools.
Figure 16. Multi-level gDT (ac) and its structure (b) are exported to gbXML (c) and revisualized as thermal zone solids (d1d3) and thematic surfaces (e1e3), demonstrating interoperability with building performance tools.
Buildings 15 04060 g016
Figure 17. Topologic Solid Model (CellComplex) with proof-of-concept simulations temperatures.
Figure 17. Topologic Solid Model (CellComplex) with proof-of-concept simulations temperatures.
Buildings 15 04060 g017
Figure 18. B-Rep model with mapped devices, simulation results, and basic consumption forecasts and analysis.
Figure 18. B-Rep model with mapped devices, simulation results, and basic consumption forecasts and analysis.
Buildings 15 04060 g018
Table 1. Summary of datasets and their key features.
Table 1. Summary of datasets and their key features.
Buildings 15 04060 i001
House 01 Block N PEA School Santa Chiara
Type of deviceMLSTLSTLSTLS
Density6 mm5mm5 mm5mm
Number of Floorplans2411
Contains indoor sensor data
Table 2. Classification results obtained using the Point Transformer v3 model.
Table 2. Classification results obtained using the Point Transformer v3 model.
ElementsStd Deviation (%)mIoU (%)mAcc (%)
Unclassified5.178.385.4
Floors2.694.296.8
Ceilings2.792.795.6
Walls2.487.589.4
Columns0.744.171.5
Doors1.151.665.3
Windows1.658.870.2
Table 3. Scan-to-BIM reconstruction metrics for classes.
Table 3. Scan-to-BIM reconstruction metrics for classes.
ElementsMetricsColumns [%]Doors [%]Walls [%]Windows [%]
House 01—First floor
NavVis dataset
mIoU70.288.668.1
F172.585.270.3
House 01—Ground Floor
NavVis dataset
mIoU85.268.989.967.8
F186.970.787.069.4
Block N
KULeuven Campus
mIoU73.491.771.3
F175.188.672.8
Santa Chiara
Trento, Italy
mIoU75.177.394.876.2
F189.178.292.077.5
School PEA
Italy
mIoU78.376.496.375.8
F181.277.994.076.7
Table 4. Metrics evaluation for both TMS- and B-Rep-based models (th = 0.065 m).
Table 4. Metrics evaluation for both TMS- and B-Rep-based models (th = 0.065 m).
ElementsTopological Solid Model (TSM)B-Rep Model (th = 0.065 m)
PrecisionRecallF1mIoUPrecisionRecallF1mIoU
House 01—Ground Floor0.9500.9030.9260.9480.8990.8790.8890.849
House 01—First Floor0.9610.9220.9410.9580.9210.9010.9110.871
Block N—Ground Floor0.9470.9300.9380.9810.9310.9110.9210.881
Block N—First Floor0.9730.9430.9580.9680.9380.9180.9280.888
Block N—Second Floor0.9840.9570.9700.9720.9460.9260.9360.896
Block N—Third Floor0.9830.9530.9680.9690.9810.9610.9710.931
Santa Chiara—Italy0.9370.8790.9070.9340.8980.8780.8880.848
School PEA—Italy0.9300.8670.8970.9270.8870.8670.8770.837
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Roman, O.; Bassier, M.; Agugiaro, G.; Arroyo Ohori, K.; Farella, E.M.; Remondino, F. Scan-to-EDTs: Automated Generation of Energy Digital Twins from 3D Point Clouds. Buildings 2025, 15, 4060. https://doi.org/10.3390/buildings15224060

AMA Style

Roman O, Bassier M, Agugiaro G, Arroyo Ohori K, Farella EM, Remondino F. Scan-to-EDTs: Automated Generation of Energy Digital Twins from 3D Point Clouds. Buildings. 2025; 15(22):4060. https://doi.org/10.3390/buildings15224060

Chicago/Turabian Style

Roman, Oscar, Maarten Bassier, Giorgio Agugiaro, Ken Arroyo Ohori, Elisa Mariarosaria Farella, and Fabio Remondino. 2025. "Scan-to-EDTs: Automated Generation of Energy Digital Twins from 3D Point Clouds" Buildings 15, no. 22: 4060. https://doi.org/10.3390/buildings15224060

APA Style

Roman, O., Bassier, M., Agugiaro, G., Arroyo Ohori, K., Farella, E. M., & Remondino, F. (2025). Scan-to-EDTs: Automated Generation of Energy Digital Twins from 3D Point Clouds. Buildings, 15(22), 4060. https://doi.org/10.3390/buildings15224060

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop