Automated Identification of Heavy BIM Library Components: A Multi-Criteria Analysis Tool for Model Optimization
Highlights
- The HeavyFamilies tool effectively identified heavy BIM components in designs of various scales (133–680 families) during an analysis lasting 8–165 s. The results confirmed that the load results from both many occurrences and complex geometry, which is why a multi-criteria assessment is needed rather than a single parameter.
- Validation involving six BIM specialists resulted in 100% task completion; automatic colour coding and direct visualisation in the model were particularly appreciated. The tool supports routine diagnostics by detecting elements with a score ≥ 200 and facilitating objective prioritisation of optimisation.
- HeavyFamilies enables a shift from reactive problem solving to proactive quality control, allowing performance to be managed early on and reducing costly corrections later. It can be implemented as a standard verification step for new components before they are accepted into the library, especially in federated and multi-disciplinary projects.
- This approach creates a universal framework for managing the quality of digital resources in BIM (e.g., completeness of information, compliance with guidelines). In the context of smart cities/CIM, it supports the transition to data-driven processes, with the potential for integration with ML (prediction of heavy components and recommendations for optimisation in the life cycle of an object).
Abstract
1. Introduction
1.1. Research Background
1.2. Research Problem
1.3. Research Gap and Purpose of the Work
- Develop a multi-criteria evaluation framework that integrates both geometric complexity metrics (solid count, face/edge count) and contextual factors (instance frequency, estimated file size contribution) to provide a comprehensive assessment of component performance impact;
- Establish a weighted scoring methodology that quantifies the relative contribution of individual library components to overall model performance, enabling objective prioritization of optimization efforts;
- Create an automated diagnostic tool integrated with industry-standard BIM software (Autodesk Revit) that eliminates the need for manual inspection and advanced technical knowledge;
- Validate the effectiveness of the proposed approach through application to real-world BIM projects of varying scales and complexity, demonstrating its practical utility in identifying optimization targets;
- Provide actionable insights to practitioners through intuitive visualization, data export capabilities, and direct model highlighting that facilitate informed decision-making in component library management.
2. Materials and Methods
2.1. Research Approach and General Assumptions
- Criteria identification: Determining relevant evaluation dimensions based on domain knowledge and empirical evidence of performance impact factors.
- Criteria measurement: Establishing quantifiable metrics that can be extracted programmatically from BIM models.
- Weight elicitation: Assigning relative importance weights to criteria based on their contribution to the decision objective (performance impact).
- Aggregation function: Combining weighted criteria into a composite score that enables component ranking and classification.
- Threshold determination: Establishing decision boundaries that separate acceptable components from those requiring intervention.
- Automation and efficiency: Model analysis must be fully automated, requiring no programming knowledge or manual parameter configuration from the user. Analysis time should be proportional to the number of family instances in the model, allowing for practical application even in large projects.
- Transparency of methodology: The user should be able to see how the weight index is calculated and interpret the results in the context of specific criteria. The tool provides detailed source data (number of instances, geometries, walls, edges) in addition to the aggregated index.
- Integration with the existing BIM ecosystem: The tool is implemented as a native plugin for the pyRevit platform, providing direct access to the Autodesk Revit API and integration with the software’s user interface. This approach eliminates the need to export data to external analytical tools.
- End-user focus: The graphical interface has been designed in accordance with user experience design principles, offering intuitive navigation, clear visualization of results, and decision-making support features (sorting, filtering, visualization in the model).
- Extensibility and documentation: Analysis results can be exported to CSV format, enabling further processing, integration with reporting systems, and the creation of performance metrics in the context of multiple projects or overtime.
2.2. Analysis Criteria and Evaluation Model
- (1)
- Empirical analysis of test projects: The tool was tested on six real BIM projects of varying scales (133–680 families per project), representing residential, commercial, and mixed-use buildings. Analysis of the Weight_Score distribution revealed a natural clustering pattern: most standard components scored below 150, while components that caused noticeable performance issues (identified through user reports and loading time measurements) consistently scored above 200. This observation suggested 200 as a meaningful boundary between acceptable and problematic components.
- (2)
- Expert consultation: The preliminary threshold was validated through consultation with six BIM specialists with 5–15 years of professional experience in architectural design offices. Experts were asked to review components from different Weight_Score ranges (50–100, 100–200, 200–300, 300+) and assess whether they would prioritize them for optimization. The consensus among experts aligned with the 200 threshold: components scoring below 200 were generally considered acceptable, while those above 200 were consistently flagged as requiring attention.
- (3)
- Performance impact correlation: In validation projects, components with Weight_Score ≥ 200 collectively accounted for a disproportionate share of model loading time and memory consumption despite representing only 5–15% of total family types. Specifically, in Project #1 (described in Section 3.2), three families exceeding the threshold (scoring 200–700+) represented less than 5% of analyzed components but contributed an estimated 25–30% of geometric processing overhead based on their combined instance count and geometric complexity.
2.3. Tool Architecture and Implementation
- (1)
- The geometry analysis module, which is responsible for extracting geometric data from family instances. It uses the Revit Geometry API to traverse the geometric hierarchy of components, identifying and counting solids, walls, and edges. The implementation includes support for nested families through recursive inspection of geometry instances (GeometryInstance). The module uses DetailLevel.Fine to ensure a complete analysis of the geometry available in the model.
- (2)
- The data aggregation module collects statistics for each unique family in the project, combining the geometric data from the first instance encountered with the family occurrence counter in the model. It uses a dictionary structure to efficiently group data by family name, ensuring O(n) computational complexity for n instances in the model. After the iteration is complete, the data is converted to FamilyData class objects that encapsulate the logic for calculating the weight index.
- (3)
- The user interface module implements a graphical interface based on Windows Forms (.NET), presenting the results in a sorted table. The interface offers row coloring functions according to severity thresholds (red for Weight_Score ≥ 200, orange for 100 < Weight_Score < 200), which increases the readability of the results and supports quick identification of problematic components. The implementation uses the DataGridView control from with configurable columns representing individual criteria and an aggregated indicator.
- (4)
- The export and visualization module provides two key functionalities: (1) exporting results to CSV format with UTF-8 BOM encoding, ensuring correct reading of Polish characters in Microsoft Excel, and (2) visualization of selected families in the model by applying graphic overrides (OverrideGraphicSettings), highlighting instances in red with bold lines. The visualization function uses Revit API transactions to modify view settings while maintaining the ability to undo changes.
2.4. User Interface and Functionalities
- (1)
- The results table (DataGridView), which is the central element of the interface, presenting all analyzed families in tabular form. The table columns represent: (1) Family Name, (2) Revit Category, (3) Number of Instances, (4) Number of Geometries, (5) Number of Faces, (6) Number of Edges, (7) estimated size (Size Est.), and (8) calculated weight score (Weight Score). The table is sorted in descending order by weight score by default, allowing the user to immediately identify the most problematic components. The user can change the sorting by clicking on the header of any column, which allows for analysis of the data from different perspectives (e.g., families with the highest number of instances, highest geometric complexity).
- (2)
- Row coloring: Table rows are automatically colored according to two severity thresholds, implementing a visual alert system. Families with Weight_Score ≥ 200 are marked with a bright red background color (RGB: 255, 200, 200), signaling a critical severity level requiring immediate attention. Families with a Weight_Score between 100 and 200 are marked with an orange background (RGB: 255, 240, 200), indicating a moderate level of severity that should be monitored. This semantic color coding supports quick visual interpretation without the need to analyze numerical values.
- (3)
- The statistics panel is located below the table and displays aggregated information in text form: “Analyzed X families | Y classified as HEAVY (weight score ≥ 200)”. This statistic provides the user with context regarding the scale of the problem in the analyzed model—the percentage of heavy families relative to the total number of unique families is a key metric for the quality of the component library.
- (1)
- Export to CSV—initiates a file save dialog, allowing the full analysis results to be exported to a CSV format with a semicolon separator and UTF-8 BOM encoding. The exported file contains all data columns visible in the table, allowing for further analysis in tools such as Microsoft Excel, Power BI, or data analysis languages (Python, R). After saving, the tool automatically opens the folder containing the exported file, optimizing the user’s workflow.
- (2)
- Highlight Selected—after selecting a row in the table and activating this function, the tool closes the dialog box and highlights all instances of the selected family in the active Revit view. The implementation uses the Selection API mechanism to select elements and OverrideGraphicSettings to apply red coloring with a weight of 5, which ensures clear visualization even in densely modeled areas. After the operation is completed, a message is displayed with the number of highlighted instances.
- (3)
- Highlight HEAVY—an advanced feature that automatically identifies all families that meet the Weight_Score ≥ 200 criterion and highlights all their instances in the model. This “big picture” tool allows the user to immediately visualize the spatial distribution of problematic components, which can reveal patterns (e.g., concentration of heavy families in specific areas of the project) that are not visible in a tabular presentation of data. A message after the operation informs about the number of highlighted families and instances.
- (4)
- Close—closes the dialog box without performing any additional operations, allowing the user to return to normal work in Revit with the option to restart the analysis later.
2.5. Validation and Testing Methodology
3. Results
3.1. Tool Installation
3.2. Heavy Families Functionality
3.3. Preview of Results in 3D View and Export to .csv
3.4. Comparative Performance Analysis Across Project Scales
4. Discussion
4.1. Interpretation of Results and Practical Implications
4.2. Limitations and Future Research Directions
4.3. Broader Context and Contribution to BIM Optimization
5. Conclusions
- Multi-criteria evaluation framework delivers measurable accuracy: The five-criteria model (weights: Instance Count 20%, Geometry Count 30%, Face Count 20%, Edge Count 10%, Estimated Size 20%) achieved 100% classification agreement with manual inspection in validation testing (n = 15 sample families), with mean relative measurement error < 4% across geometric criteria. This demonstrates that multi-criteria aggregation is not only theoretically sound but empirically reliable.
- Threshold-based classification is robust across project scales: Validation on three real-world projects (133, 240, 680 families; residential, office, mixed-use) showed consistent identification of 1.5–3.4% heavy components regardless of project size—representing 2–23 absolute components per project. The Weight Score threshold of 200 proved stable across 5-fold scale variation, with maximum scores ranging 287–734 depending on project complexity.
- Automated analysis achieves significant efficiency gains: Processing time scaled linearly (8 s for 133 families → 165 s for 680 families; O(n) complexity confirmed), representing >95% time reduction compared to manual inspection (estimated hours vs. seconds). Tool analysis identified components with Weight Scores ranging from dozens to 700+ points, with automatic color-coding enabling instant prioritization.
- Practical validation confirms systematic improvement over current practice: User testing with six BIM specialists (5–15 years experience) achieved 100% task completion rate. The tool identified problematic components (Weight Score ≥ 200) that were missed by manual inspection, demonstrating that automated multi-criteria analysis detects performance issues not evident through subjective assessment.
- The contribution bridges scientific methodology and industrial application: Scientifically, the work establishes a generalizable MCDA framework for BIM component quality assessment with empirically validated criteria weights and thresholds. Industrially, the open-source pyRevit implementation provides immediate practical utility for performance diagnostics, library quality control, and component comparison in real design workflows.
Funding
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Sacks, R.; Eastman, C.; Lee, G.; Teicholz, P. BIM Handbook: A Guide to Building Information Modeling for Owners, Designers, Engineers, Contractors, and Facility Managers, 3rd ed.; John Wiley & Sons: Hoboken, NJ, USA, 2018. [Google Scholar]
- Azhar, S. Building Information Modeling (BIM): Trends, Benefits, Risks, and Challenges for the AEC Industry. Leadersh. Manag. Eng. 2011, 11, 241–252. [Google Scholar] [CrossRef]
- Volk, R.; Stengel, J.; Schultmann, F. Building Information Modeling (BIM) for existing buildings—Literature review and future needs. Autom. Constr. 2014, 38, 109–127. [Google Scholar] [CrossRef]
- Borkowski, A.S. File Hygiene and BIM Models Restrictions. Trends Civ. Eng. Its Archit. 2019, 3, 430–435. [Google Scholar] [CrossRef]
- Gu, N.; London, K. Understanding and facilitating BIM adoption in the AEC industry. Autom. Constr. 2010, 19, 988–999. [Google Scholar] [CrossRef]
- Singh, V.; Gu, N.; Wang, X. A theoretical framework of a BIM-based multi-disciplinary collaboration platform. Autom. Constr. 2011, 20, 134–144. [Google Scholar] [CrossRef]
- Liu, Y.; van Nederveen, S.; Hertogh, M. Understanding effects of BIM on collaborative design and construction: An empirical study in China. Int. J. Proj. Manag. 2017, 35, 686–698. [Google Scholar] [CrossRef]
- Porwal, A.; Hewage, K.N. Building Information Modeling (BIM) partnering framework for public construction projects. Autom. Constr. 2013, 31, 204–214. [Google Scholar] [CrossRef]
- Kreider, R.G.; Messner, J.I. The Uses of BIM: Classifying and Selecting BIM Uses; Penn State Computer Integrated Construction Research Program; The Pennsylvania State University: University Park, PA, USA, 2013; Version 0.9. [Google Scholar]
- NBS. National BIM Report 2021; RIBA Enterprises Ltd.: Newcastle Upon Tyne, UK, 2021. [Google Scholar]
- Borrmann, A.; König, M.; Koch, C.; Beetz, J. Building Information Modeling: Technology Foundations and Industry Practice; Springer: Cham, Switzerland, 2018. [Google Scholar]
- Whyte, J.; Stasis, A.; Lindkvist, C. Managing change in the delivery of complex projects: Configuration management, asset information and ‘big data’. Int. J. Proj. Manag. 2016, 34, 339–351. [Google Scholar] [CrossRef]
- Hu, Z.Z.; Tian, P.L.; Li, S.W.; Zhang, J.P. BIM-based integrated delivery technologies for intelligent MEP management in the operation and maintenance phase. Adv. Eng. Softw. 2018, 115, 1–16. [Google Scholar] [CrossRef]
- BIMForum. Level of Development (LOD) Specification Part I & Commentary; BIMForum: Washington, DC, USA, 2021. [Google Scholar]
- AIA. AIA Document G202–2013: Project Building Information Modeling Protocol Form; The American Institute of Architects: Washington, DC, USA, 2013. [Google Scholar]
- ISO 19650-1:2018; Organization and Digitization of Information about Buildings and Civil Engineering Works, Including Building Information Modeling (BIM)—Information Management Using Building Information Modeling. International Organization for Standardization: Geneva, Switzerland, 2018.
- Autodesk Inc. Revit Performance Technical Note; Autodesk Knowledge Network: San Rafael, CA, USA, 2023. [Google Scholar]
- Oti, A.H.; Tizani, W. BIM extension for the sustainability appraisal of conceptual steel design. Adv. Eng. Inform. 2015, 29, 28–46. [Google Scholar] [CrossRef]
- Cavka, H.B.; Staub-French, S.; Pottinger, R. Evaluating the alignment of organizational and project contexts for BIM adoption: A case study of a large owner organization. Buildings 2015, 5, 1265–1300. [Google Scholar] [CrossRef]
- Daum, S.; Borrmann, A. Processing of Topological BIM Queries using Boundary Representation Based Methods. Adv. Eng. Inform. 2014, 28, 272–286. [Google Scholar] [CrossRef]
- Won, J.; Lee, G.; Dossick, C.; Messner, J. Where to focus for successful adoption of building information modeling within organization. J. Constr. Eng. Manag. 2013, 139, 04013018. [Google Scholar] [CrossRef]
- Motawa, I.; Almarshad, A. A knowledge-based BIM system for building maintenance. Autom. Constr. 2013, 29, 173–182. [Google Scholar] [CrossRef]
- Baryłka, A.; Ogrodnik, P.; Żółtkowski, M.; Obolewicz, J.; Tomaszewicz, D. The construction process in the era of digitalization of construction. Mod. Eng. 2025, 3, 34–38. [Google Scholar]
- Lee, P.-C.; Wang, Y.; Lo, T.-P.; Long, D. An Integrated System Framework of Building Information Modelling and Geographical Information System for Utility Tunnel Maintenance Management. Appl. Sci. 2018, 8, 2117. [Google Scholar] [CrossRef]
- Matthews, J.; Love, P.E.D.; Heinemann, S.; Chandler, R.; Rumsey, C.; Olatunj, O. Real time progress management: Re-engineering processes for cloud-based BIM in construction. Autom. Constr. 2015, 58, 38–47. [Google Scholar] [CrossRef]
- Borkowski, A.S. Introduction to BIM: Philosophy of Building Information Modeling; Warsaw University of Technology Press: Warsaw, Poland, 2024; 188p. (In Polish) [Google Scholar]
- Leite, F.; Akcamete, A.; Akinci, B.; Atasoy, G.; Kiziltas, S. Analysis of modeling effort and impact of different levels of detail in building information models. Autom. Constr. 2011, 20, 601–609. [Google Scholar] [CrossRef]
- Grilo, A.; Jardim-Goncalves, R. Value proposition on interoperability of BIM and collaborative working environments. Autom. Constr. 2010, 19, 522–530. [Google Scholar] [CrossRef]
- Hardin, B.; McCool, D. BIM and Construction Management: Proven Tools, Methods, and Workflows, 2nd ed.; John Wiley & Sons: Indianapolis, IN, USA, 2015. [Google Scholar]
- Kim, H.; Anderson, K.; Lee, S.; Hildreth, J. Generating construction schedules through automatic data extraction using open BIM (building information modeling) technology. Autom. Constr. 2013, 35, 285–295. [Google Scholar] [CrossRef]
- Belton, V.; Stewart, T. Multiple Criteria Decision Analysis: An Integrated Approach; Springer Science & Business Media: Boston, MA, USA, 2002. [Google Scholar]
- Thokala, P.; Devlin, N.; Marsh, K.; Baltussen, R.; Boysen, M.; Kalo, Z.; Longrenn, T.; Mussen, F.; Peacock, S.; Watkins, J.; et al. Multiple Criteria Decision Analysis for Health Care Decision Making—An Introduction: Report 1 of the ISPOR MCDA Emerging Good Practices Task Force. Value Health 2016, 19, 1–13. [Google Scholar] [CrossRef] [PubMed]
- Hevner, A.R.; March, S.T.; Park, J.; Ram, S. Design Science in Information Systems Research. MIS Q. 2004, 28, 75–105. [Google Scholar] [CrossRef]
- Jiang, S.; Jiang, L.; Han, Y.; Wu, Z.; Wang, N. OpenBIM: An Enabling Solution for Information Interoperability. Appl. Sci. 2019, 9, 5358. [Google Scholar] [CrossRef]
- Iran-Nejad, E. pyRevit: Rapid Application Development (RAD) Environment for Autodesk Revit. Available online: https://github.com/pyrevitlabs/pyRevit (accessed on 10 December 2024).
- Brooke, J. SUS: A “Quick and Dirty” Usability Scale. In Usability Evaluation in Industry; Jordan, P.W., Thomas, B., Weerdmeester, B.A., McClelland, I.L., Eds.; Taylor & Francis: London, UK, 1996; pp. 189–194. [Google Scholar]
- Zhang, C.; Arditi, D. Automated progress control using laser scanning technology. Autom. Constr. 2013, 36, 108–116. [Google Scholar] [CrossRef]
- Kang, T.W.; Hong, C.H. A study on software architecture for effective BIM/GIS-based facility management data integration. Autom. Constr. 2015, 54, 25–38. [Google Scholar] [CrossRef]
- Deng, Y.; Cheng, J.C.; Anumba, C. Mapping between BIM and 3D GIS in different levels of detail using schema mediation and instance comparison. Autom. Constr. 2016, 67, 1–21. [Google Scholar] [CrossRef]
- Xu, X.; Ma, L.; Ding, L. A framework for BIM-enabled life-cycle information management of construction project. Int. J. Adv. Robot. Syst. 2014, 11, 126. [Google Scholar] [CrossRef]
- Radziejowska, A.; Ciepłucha, W.; Majta, M. BIM model for the operational phase based on available documentation. Arch. Civ. Eng. 2025, 71, 347–363. [Google Scholar] [CrossRef]
- Radziejowska, A.; Ciepłucha, W.; Majta, M. Pilot implementation of a digital building model for operational management. Arch. Civ. Eng. 2025, 71, 365–380. [Google Scholar] [CrossRef]
- Kim, J.; Kent, M.; Kral, K.; Dogan, T. Seemo: A new tool for early design window view satisfaction evaluation in residential buildings. Build. Environ. 2022, 214, 108909. [Google Scholar] [CrossRef]






| Component Category | Instance Count (Typical) | Geometry Count | Face Count | Edge Count | Weight Score (Typical Range) |
|---|---|---|---|---|---|
| Simple furniture (chair, table) | 10–50 | 5–20 | 20–100 | 50–200 | 50–150 |
| Standard doors/windows | 20–100 | 10–30 | 50–200 | 100–500 | 80–200 |
| MEP components (ducts, pipes) | 50–500 | 3–15 | 10–80 | 30–150 | 100–250 |
| Complex fixtures (sinks, toilets) | 5–30 | 20–100 | 100–500 | 200–1000 | 150–400 |
| Detailed architectural elements | 10–50 | 50–200 | 300–1500 | 500–3000 | 250–600+ |
| Parameter | Project A (Small) | Project B (Medium) | Project C (Large) |
|---|---|---|---|
| Building type | Residential (single-family) | Office building | Mixed-use complex |
| Total families | 133 | 240 | 680 |
| Heavy families (≥200) | 2 (1.5%) | 8 (3.3%) | 23 (3.4%) |
| Max Weight Score | 287 | 456 | 734 |
| Avg Weight Score (all) | 42 | 68 | 81 |
| Avg Weight Score (heavy) | 243 | 289 | 337 |
| Analysis time | 8 s | 35 s | 165 s |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2026 by the author. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license.
Share and Cite
Borkowski, A.S. Automated Identification of Heavy BIM Library Components: A Multi-Criteria Analysis Tool for Model Optimization. Smart Cities 2026, 9, 22. https://doi.org/10.3390/smartcities9020022
Borkowski AS. Automated Identification of Heavy BIM Library Components: A Multi-Criteria Analysis Tool for Model Optimization. Smart Cities. 2026; 9(2):22. https://doi.org/10.3390/smartcities9020022
Chicago/Turabian StyleBorkowski, Andrzej Szymon. 2026. "Automated Identification of Heavy BIM Library Components: A Multi-Criteria Analysis Tool for Model Optimization" Smart Cities 9, no. 2: 22. https://doi.org/10.3390/smartcities9020022
APA StyleBorkowski, A. S. (2026). Automated Identification of Heavy BIM Library Components: A Multi-Criteria Analysis Tool for Model Optimization. Smart Cities, 9(2), 22. https://doi.org/10.3390/smartcities9020022

