Large Language Models for Structured Information Processing in Construction and Facility Management
Abstract
1. Introduction
1.1. Problem Statement
1.2. Thesis Statement and Research Questions
- Unstructured Input + Unstructured Output: A user submits a free-text question and the LLM responds with a narrative explanation based on retrieved manuals or reports. For instance, a user might ask for the types of floor covers (e.g., for planning cleaning operations), expecting a text answer providing further details.
- Unstructured Input + Structured Output: In this scenario, the user submits a natural language query and the LLM responds by generating a structured output, such as an SQL query. This is a use case to support a scenario where BIM was already used involving a database with relevant information. For instance, a user might ask for all the cold storage rooms with their areas.
- Structured Input + Structured Output: The user provides structured parameters and the LLM generates a corresponding structured response, such as a command, database query, or summarized data output, to support automated workflows. This is an advanced scenario with a human user familiar with structured information access, e.g., in the form of SQL queries. It could also include automated information access by an application, e.g., a software for planning operations.
- Use Case 1 considers only the first step from a user query in natural language involving the LLM-based user interface to a structured output assuming to be a suitable SQL query.
- Use Case 2 goes one step further and includes the evaluation of the SQL query by the related database software.
- Use Case 3 considers the end-to-end process by including the evaluation and presentation of the SQL query result through the LLM.
- MRQ: Assuming a scenario based on using information from a relational database, how does the integration of unstructured/structured data impact the retrieval accuracy of LLM-based facility management systems?
- SRQ1: What specific parameters (such as for the prompt specification) are relevant and most effective for handling and extracting insights from structured facility management data?
- SRQ2: What are the advantages and limitations of different methods/techniques for providing structured or unstructured input and output to LLMs in terms of processing efficiency and response correctness?
- SRQ3: How do different structured and unstructured input strategies (e.g., structured queries, database schema, guided prompts) affect the ability of LLMs to retrieve and process information in a facility management context?
- SRQ4: What factors (e.g., input/output formats, LLM model choice, retrieval technique) most significantly influence the success of structured-data-enhanced LLM applications in real-world scenarios?
1.3. Scope
- Understand domain-specific natural language queries;
- Generate syntactically and semantically valid SQL statements;
- Correctly reflect user intent based on the database schema;
- Support complex filtering, aggregation, and conditional logic;
- Communicate results back to the user in an accurate and context-aware manner.
2. Literature Review
2.1. Literature Search
2.2. Information Management in Facility Management
2.3. Enhancing LLMs with Unstructured or Structured Data
2.4. Summary and Research Gap
3. Research Methodology
3.1. Problem Awareness
3.2. Suggested Solution
- Prompt engineering with different configurations (zero-shot, few-shot, and guided prompts);
- RAG;
- Function calling and structured output (e.g., SQL, JSON) generation;
- Embedding-based document search, including the use of multilingual models such as BGE-M3.
- Input Format: unstructured queries vs. structured parameter input;
- Output Format: free-text explanations vs. structured SQL/JSON output;
- Data Type: unstructured document corpus (e.g., reports, manuals) vs. structured relational data (PostgreSQL);
- Retrieval Strategy: zero-context vs. embedding-based RAG with semantic chunking.
3.3. Data Sources
- Operation and maintenance manuals (PDF);
- Building and infrastructure plans;
- Environmental reports and concept papers;
- Financial summaries related to the building operation.
- Room and asset metadata;
- Maintenance schedules;
- Inspection logs;
- Equipment classifications.
3.4. Experimental Evaluation
- Correctness: factual accuracy based on comparison with database or document ground truth;
- Hallucination: degree to which fabricated or irrelevant information was included;
- Contextual Relevance: semantic match between the query intent and output;
- Execution Validity: the syntactical and semantically correctness of the structured outputs (SQL queries).
- A comparative results matrix for all tested use case configurations;
- Annotated examples of the outputs;
- Taxonomy of input/output structures and their impact on model behavior;
- Practical guidelines for prompt design and structured data preparation in facility management LLM deployments.
4. Experiment Design and Implementation
4.1. Integration with Facility Management Data
4.2. Setup and Tools
4.3. Experiment Design
4.3.1. Use Case 1: Unstructured Input → Structured Output
- Example Input: “What is the area of Room 203 in Building A?”
- Expected Output: SELECT area FROM rooms WHERE room_number = ‘203’ AND building_id = ‘A’.
4.3.2. Use Case 2: Unstructured Input → Function Calling
- Example Input: “Calculate the total floor area of Building A using the Libal platform.”
- Example Output: CALL calculate_floor_area(‘Building A’) → returns 1254.67 m2.
4.3.3. Use Case 3: Structured Input → Unstructured Output
- Example Input: prompt = (f’Question: On which floors are training rooms available?n’;
- f’SQL Query: SELECT id, name FROM space WHERE description_txt LIKE ‘%training%’;\n’;
- f’SQL Result: [(74592546, ‘00.415’), (74622500, ‘01.127’), (74632504, ‘01.535’)]’);
- Example Output: The rooms with IDs 74592546, 74622500, and 74632504, named ‘00.415’, ‘01.127’, and ‘01.535’, are designated for training purposes.
4.4. Prompt Strategies
- Simple prompts (easy)
- “On which floors are restrooms located?”
- “What types of floor coverings exist?”
- 2.
- Medium-complexity prompts (medium)
- “List all cooling rooms with their area sizes.”
- “What are the area sizes of the sanitary rooms on each floor?”
- 3.
- Complex prompts (hard)
- “Which training room has the largest area?”
- “Which rooms are suitable for training?”
4.5. Evaluation Criteria
- Correctness: measures whether the generated SQL query produces the correct results based on the known ground truth from the PostgreSQL database.
- Hallucination: evaluates the degree to which the model includes fabricated, irrelevant, or unsupported information in its response.
- Contextual relevance: checks how well the generated output aligns with the user’s original query intent and the relevant database schema.
- Execution validity: verifies whether the generated SQL queries are syntactically correct and can be executed without error on the PostgreSQL database. This criterion ensures that the model output is not only logically plausible but also technically usable in real-world systems.
5. Implementation Architecture
5.1. Text-to-Query-to-Text Pipeline
- User query translation (write_query): The LLM receives a user prompt in either an unstructured or structured form and generates a corresponding SQL query. This step tests the model’s ability to understand the intent and develop the correct logic. This phase corresponds to the defined use case in the following section: Use Case 1: Unstructured Input → Structured Output.
- Query execution (execute_query): The generated SQL query is executed against a PostgreSQL database containing facility management data. This step ensures that the SQL is valid, executable, and returns meaningful results.
- Response generation (generate_answer): The result is post-processed and returned as a natural language unstructured or structured response, depending on the test configuration. This phase corresponds to the defined Use Case 3: Structured Input → Unstructured Output.
5.2. Setup for the Experiments
5.2.1. Test Questions
5.2.2. Evaluation Setup and Rules
6. Results
- Correctness (alignment with the expected answers);
- Execution validity (syntax and logic of the generated SQL);
- Hallucination (presence of false or fabricated content);
- Contextual relevance (semantic fit to the prompt).
6.1. Key Findings
- Execution validity (share of queries that run without error) rises from 0% without the schema to 35% with the schema. Without an explicit context, every generated query in our test set failed to execute.
- Correctness (at least one of the expected answer values is present in the SQL or the natural-language answer) improves from 38% to 58% of all runs when the schema is supplied.
- Hallucinations (answers containing no meaningful overlap with expected values) dropped from 65% without schema to 50% with schema, which is a 15-point reduction but still indicating that half of the schema-aware answers lacked grounding.
6.2. Additional Findings
6.3. Observed Hallucinations
- Fictitious room labels: Some models invented room numbers or names that did not exist in the dataset.
- Incorrect attribute assignments: In several responses, materials or room functions were mentioned that did not match the actual query or database content.
- Unjustified generalizations: In some cases, the models generated broad statements such as “There are several rooms with parquet flooring,” even though such information was not supported by the retrieved context.
6.4. Contextual Relevance
7. Discussion and Conclusions
- Importance of schema-based prompting: Providing explicit access to database schema information significantly increases execution validity and correctness. LLMs frequently generate invalid or semantically incorrect queries without a structured context. Therefore, table definitions or attribute structures should be embedded directly in prompt templates, especially for tasks involving automated data retrieval.
- Structured inputs improve robustness: Structured inputs—such as JSON-based queries or parameterized prompts—substantially enhance consistency and reduce hallucinations. Input formalization should be prioritized in practice, particularly for tasks such as space allocation, asset reporting, or cleaning logistics.
- Prompt engineering as a critical skill: Although not the focus of our study, the experiments showed that the design and formulation of prompts directly affect the quality of LLM outputs. Task-specific, structured prompts aligned with the database schema and user intent are crucial. Organizations should develop internal guidelines for prompt design or provide tools to support structured prompt generation.
- Model selection based on contextual needs: Performance differences between the models were significant. Advanced models such as Qwen, Gemini 2.5 Flash, and LLaMA 3.3 70B demonstrated stronger performance in SQL generation and contextual understanding. Model selection should be based on empirical benchmarking that reflects specific use case requirements and available structured input.
- Integration into existing systems: The evaluated text-to-SQL-to-text pipeline is suitable for integration into facility management platforms via APIs or modular agent architectures. LLM-based assistants can be embedded to translate user queries into SQL and return natural language summaries—eliminating the need for manual database interaction.
- Data privacy and system architecture: Facility management data often include sensitive information. Locally hosted LLMs or secure hybrid deployments are preferred over public APIs. Measures such as data minimization, encryption, and internal compliance with privacy regulations should be standard.
- Use of hybrid architectures: Combining LLMs with rule-based validation and RAG can increase reliability and safety. Schema metadata can be retrieved dynamically and added to prompts, while rule-based systems validate the generated queries before execution.
- Schema completeness: Only selected tables were included in the schema because of token limitations. In future studies, larger-context models such as Gemini 1m could be tested with the full database schema to assess performance improvements. In addition, there should be further research on systematically reducing BIM models (including database schema) to make the models more appropriate for applications in facility management.
- Prompt diversity and reusability: More prompt variations could be introduced, including multilingual input and more abstract queries. This would test the model’s adaptability and robustness.
- Test sample size: For future work, we suggest more comprehensive studies involving a larger set of test questions which should provide more reliability in the results and allow for significance testing.
- Bias in evaluation: Some evaluation steps used automated tools or LLM-as-a-judge methods. Although helpful, they can introduce bias. Using the manual screening of our results, we did not find any related issues, which was probably because of the availability of a ground truth and the clear definition of evaluation criteria. However, a more diverse human evaluation could help ensure fairness and improve the quality of the results.
- Consistency tests: multiple runs with the same prompt could be used to test the output stability (assuming a temperature parameter > 0). This would help us to better understand the model reliability in productive settings.
- Expand to RAG and function calling: While RAG based on using documents was excluded in this study, combining a text-to-SQL function calling with RAG may leverage further information not included in the database but in various documents (as common in construction and facility management). This could allow for more advanced applications in these industries.
- Long-term integration testing: Future research could include real-time tests in operational systems to evaluate latency, user experience, and scalability under practical conditions.
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Lindkvist, C.; Whyte, J. Challenges and opportunities involving facilities management in data handover: London 2012 case study. In Proceedings of the AEI 2013: Building Solutions for Architectural Engineering, State College, PA, USA, 3–5 April 2013; American Society of Civil Engineers: Reston, VA, USA, 2013; pp. 670–679. [Google Scholar] [CrossRef]
- Krütli, D.; Hanne, T. Augmenting LLMs to Securely Retrieve Information for Construction and Facility Management. Information 2025, 16, 76. [Google Scholar] [CrossRef]
- William East, E.; Nisbet, N.; Liebich, T. Facility management handover model view. J. Comput. Civ. Eng. 2013, 27, 61–67. [Google Scholar] [CrossRef]
- Pinti, L.; Bonelli, S. A methodological framework to optimize data management costs and the Hand-Over phase in cultural heritage projects. Buildings 2022, 12, 1360. [Google Scholar] [CrossRef]
- Hosseini, M.R.; Roelvink, R.; Papadonikolaki, E.; Edwards, D.J.; Pärn, E. Integrating BIM into facility management: Typology matrix of information handover requirements. Int. J. Build. Pathol. Adapt. 2018, 36, 2–14. [Google Scholar] [CrossRef]
- Zhu, L.; Shan, M.; Xu, Z. Critical review of building handover-related research in construction and facility management journals. Eng. Constr. Archit. Manag. 2021, 28, 154–173. [Google Scholar] [CrossRef]
- Abdelkarim, S.B.; Ahmad, A.M.; Naji, K. A BIM-based framework for managing handover information loss. J. Manag. Eng. 2024, 40, 04024030. [Google Scholar] [CrossRef]
- Gao, Y.; Xiong, Y.; Gao, X.; Jia, K.; Pan, J.; Bi, Y.; Dai, Y.; Sun, J.; Wang, M.; Wang, H. Retrieval-Augmented Generation for Large Language Models: A Survey. arXiv 2024, arXiv:2312.10997. [Google Scholar] [CrossRef]
- Atkin, B.; Brooks, A. Total Facility Management, 5th ed.; John Wiley & Sons: Hoboken, NJ, USA, 2021. [Google Scholar]
- Dixit, M.K.; Venkatraj, V.; Ostadalimakhmalbaf, M.; Pariafsai, F.; Lavy, S. Integration of facility management and building information modeling (BIM) A review of key issues and challenges. Facilities 2019, 37, 455–483. [Google Scholar] [CrossRef]
- Matarneh, S.T.; Danso-Amoako, M.; Al-Bizri, S.; Gaterell, M.; Matarneh, R. Building information modeling for facilities management: A literature review and future research directions. J. Build. Eng. 2019, 24, 100755. [Google Scholar] [CrossRef]
- Altohami, A.B.A.; Haron, N.A.; Ales Alias, A.H.; Law, T.H. Investigating approaches of integrating BIM, IoT, and facility management for renovating existing buildings: A review. Sustainability 2021, 13, 3930. [Google Scholar] [CrossRef]
- Guo, H.; Zhou, Y.; Ye, X.; Luo, Z.; Xue, F. Automated mapping from an IFC data model to a relational database model. J. Tsinghua Univ. 2020, 61, 152–160. [Google Scholar]
- Barzegar, M.; Rajabifard, A.; Kalantari, M.; Atazadeh, B. An IFC-based database schema for mapping BIM data into a 3D spatially enabled land administration database. Int. J. Digit. Earth 2021, 14, 736–765. [Google Scholar] [CrossRef]
- Dai, C.; Cheng, K.; Liang, B.; Zhang, X.; Liu, Q.; Kuang, Z. Digital twin modeling method based on IFC standards for building construction processes. Front. Energy Res. 2024, 12, 1334192. [Google Scholar] [CrossRef]
- Toldo, B.M.; Modolo, A.; Zanchetta, C.; Bock, B.S. SQL Relational Database Usage for Integration Between BIM Models and Quantity Take-Off Platforms. In Proceedings of the New Frontiers of Construction Management (CMW 2024), Ravenna, Italy, 7–8 November 2024; Construction Management Workshop. Springer Nature: Cham, Switzerland, 2024; pp. 165–176. [Google Scholar]
- Pidgeon, A.; Dawood, N. BIM adoption issues in infrastructure construction projects: Analysis and solutions. J. Inf. Technol. Constr. 2021, 26, 263–285. [Google Scholar] [CrossRef]
- Pinti, L.; Codinhoto, R.; Bonelli, S. A review of building information modelling (BIM) for facility management (FM): Implementation in public organisations. Appl. Sci. 2022, 12, 1540. [Google Scholar] [CrossRef]
- Masataka, H.; Yutaka, W. Making software based on human-driven design case study: SQL for non-experts. In Proceedings of the 2022 IEEE 15th International Symposium on Embedded Multicore/Many-Core Systems-on-Chip (MCSoC), Penang, Malaysia, 19–22 December 2022; IEEE: New York, NY, USA, 2022; pp. 264–270. [Google Scholar] [CrossRef]
- Taipalus, T. The effects of database complexity on SQL query formulation. J. Syst. Softw. 2020, 165, 110576. [Google Scholar] [CrossRef]
- Baig, M.S.; Imran, A.; Yasin, A.U.; Butt, A.H.; Khan, M.I. Natural language to SQL queries: A review. Int. J. Innov. Sci. Technol. 2022, 4, 147–162. [Google Scholar] [CrossRef]
- Shorten, C.; Pierse, C.; Smith, T.B.; D’Oosterlinck, K.; Celik, T.; Cardenas, E.; Monigatti, L.; Hasan, M.S.; Schmuhl, E.; Williams, D.; et al. Querying Databases with Function Calling. arXiv 2025, arXiv:2502.00032. [Google Scholar] [CrossRef]
- Ji, Z.; Lee, N.; Frieske, R.; Yu, T.; Su, D.; Xu, Y.; Ishii, E.; Bang, Y.J.; Madotto, A.; Fung, P. Survey of Hallucination in Natural Language Generation. ACM Comput. Surv. 2023, 55, 1–38. [Google Scholar] [CrossRef]
- Dettmers, T.; Pagnoni, A.; Holtzman, A.; Zettlemoyer, L. QLoRA: Efficient Finetuning of Quantized LLMs. arXiv 2023, arXiv:2305.14314. [Google Scholar] [CrossRef]
- Lewis, P.; Perez, E.; Piktus, A.; Petroni, F.; Karpukhin, V.; Goyal, N.; Küttler, H.; Lewis, M.; Yih, W.; Rocktäschel, T.; et al. Retrieval-augmented generation for knowledge-intensive NLP tasks. Adv. Neural Inf. Process. Syst. 2020, 33, 9459–9474. Available online: https://proceedings.neurips.cc/paper/2020/file/6b493230205f780e1bc26945df7481e5-Paper.pdf (accessed on 2 July 2025).
- Borgeaud, S.; Mensch, A.; Hoffmann, J.; Cai, T.; Rutherford, E.; Millican, K.; Driessche, G.B.V.D.; Lespiau, J.-B.; Damoc, B.; Clark, A.; et al. Improving Language Models by Retrieving from Trillions of Tokens. In Proceedings of the 39th International Conference on Machine Learning, Baltimore, MD, USA, 17–23 June 2022; Microtome Publishing: Brookline, MA, USA, 2022; pp. 2206–2240. Available online: https://proceedings.mlr.press/v162/borgeaud22a.html (accessed on 2 July 2025).
- Liu, X.; Sun, J.; Lei, A.; Zhu, J. Research and Applications of Large Language Models for Converting Unstructured Data into Structured Data. In Proceedings of the 2024 3rd International Conference on Cloud Computing, Big Data Application and Software Engineering (CBASE), Hangzhou, China, 11–13 October 2024; IEEE: New York, NY, USA, 2024; pp. 305–308. [Google Scholar] [CrossRef]
- Gim, I.; Lee, S.; Zhong, L. Asynchronous LLM Function Calling. arXiv 2024, arXiv:2412.07017. [Google Scholar] [CrossRef]
- Irugalbandara, C. Meaning Typed Prompting: A Technique for Efficient, Reliable Structured Output Generation. arXiv 2024, arXiv:2410.18146. [Google Scholar] [CrossRef]
- Finegan-Dollak, C.; Kummerfeld, J.K.; Zhang, L.; Ramanathan, K.; Sadasivam, S.; Zhang, R.; Radev, D. Improving text-to-sql evaluation methodology. arXiv 2018, arXiv:1806.09029. [Google Scholar] [CrossRef]
- Katsogiannis-Meimarakis, G.; Koutrika, G. A survey on deep learning approaches for text-to-SQL. VLDB J. 2023, 32, 905–936. [Google Scholar] [CrossRef]
- Yu, T.; Zhang, R.; Yang, K.; Yasunaga, M.; Wang, D.; Li, Z.; Ma, J.; Li, I.; Yao, Q.; Roman, S.; et al. Spider: A large-scale human-labeled dataset for complex and cross-domain semantic parsing and text-to-SQL task. arXiv 2018, arXiv:1809.08887. [Google Scholar] [CrossRef]
- Zhong, V.; Xiong, C.; Socher, R. Seq2sql: Generating structured queries from natural language using reinforcement learning. arXiv 2017, arXiv:1709.00103. [Google Scholar] [CrossRef]
- Mitsopoulou, A.; Koutrika, G. Analysis of text-to-SQL benchmarks: Limitations, challenges and opportunities. In Proceedings of the 28th International Conference on Extending Database Technology, EDBT, Barcelona, Spain, 25–28 March 2025; OpenProceedings.org: Konstanz, Germany, 2025; pp. 199–212. Available online: https://datagems.eu/wp-content/uploads/2025/05/paper-41.pdf (accessed on 2 July 2025).
- Liu, X.; Shen, S.; Li, B.; Ma, P.; Jiang, R.; Zhang, Y.; Fan, J.; Li, G.; Tang, N.; Luo, Y. A Survey of Text-to-SQL in the Era of LLMs: Where are we, and where are we going? IEEE Trans. Knowl. Data Eng. 2025, 37, 5735–5754. [Google Scholar] [CrossRef]
- Hong, Z.; Yuan, Z.; Zhang, Q.; Chen, H.; Dong, J.; Huang, F.; Huang, X. Next-generation database interfaces: A survey of LLM-based text-to-SQL. IEEE Trans. Knowl. Data Eng. 2025, 1–20. [Google Scholar] [CrossRef]
- Chen, J.; Xiao, S.; Zhang, P.; Luo, K.; Lian, D.; Liu, Z. BGE M3-Embedding: Multi-Lingual, Multi-Functionality, Multi-Granularity Text Embeddings Through Self-Knowledge Distillation. arXiv 2024, arXiv:2402.03216. [Google Scholar] [CrossRef]
- Chung, J.; Gulcehre, C.; Cho, K.; Bengio, Y. Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling. arXiv 2014, arXiv:1412.3555. [Google Scholar] [CrossRef]
- Günther, M.; Ong, J.; Mohr, I.; Abdessalem, A.; Abel, T.; Akram, M.K.; Guzman, S.; Mastrapas, G.; Sturua, S.; Wang, B.; et al. Jina Embeddings 2: 8192-Token General-Purpose Text Embeddings for Long Documents). arXiv 2024, arXiv:2310.19923. [Google Scholar] [CrossRef]
- Hevner, A.R.; March, S.T.; Park, J.; Ram, S. Design science in information systems research. MIS Q. 2004, 28, 75–105. [Google Scholar] [CrossRef]
- Peffers, K.; Tuunanen, T.; Rothenberger, M.A.; Chatterjee, S. A design science research methodology for information systems research. J. Manag. Inf. Syst. 2007, 24, 45–77. [Google Scholar] [CrossRef]
- CDE für Bauherren und Betreiber. LIBAL. Available online: https://www.libal-tech.de/libal-common-data-environment-cde/ (accessed on 3 June 2025).
Component | Description |
---|---|
Evaluated LLMs | Gemini (Google): gemini-1.5-flash gemini-2.0-flash gemini-2.5-flash-preview-04-17 gemini-2.5-flash-preview-05-20 Llama (Meta): llama-3.1-8b llama-3.3-70b llama-4-scout-17b-16e-instruct llama-4-maverick-17b-128e-instruct Qwen (Alibaba Cloud): qwen-3-32b Mistral (Mistral AI): mistral-saba-24b |
LLM Inference Providers | Cerebras—Llama 3.1, 3.3, Scout and Qwen models Gemini API—all Gemini models Groq—Mistral and Llama 4 Maverick models |
AI Frameworks | LangChain (structured/unstructured input/output for writing the query, executing it, and generating the final answer for all listed LLMs) |
Data Sources | PostgreSQL Libal database (structured data), PDFs (manuals, plans), Libal API |
Development Tools | Python 3.12, Jupyter Notebooks 7.4, PyCharm 2024.3, vs. Code 1.97, GitHub |
Questions | Expected Output | Intention | Expected Information Source |
---|---|---|---|
On which floors are restrooms available? | 1F, 2F, 3F, 4F, 5F, 6F, 7F, 8F | Planning cleaning effort | LIBAL database |
What types of floor covers are there? | Stone and cement floors, stone and tile floors, plastic coverings and linoleum, parquet and cork parquet, unknown, mastic asphalt and rubber floors, textile flooring | Determination of the floor covers | LIBAL database |
Give me all the cold storage rooms with their areas. | Cold storage room 00.631 has a total area of 6.299 m2, cold storage room 00.632 has a total area of 6.299 m2, cold storage room 00.636 has a total area of 5.098 m2, cold storage room 00.637 has a total area of 8.075 m2, cold storage room 00.638 has a total area of 22 m2, cold storage room 00.645 has a total area of 7.234 m2 | Survey of the storage area | LIBAL database |
What are the areas of the individual restrooms? | B1 = 0 m2, 1F = 118.65 m2, 2F = 39.14 m2, 3F = 39.14 m2, 4F = 39.14 m2, 5F = 39.14 m2, 6F = 39.14 m2, 7F = 39.14 m2, 8F = 39.14 m2, 9F = 39.14 m2, 10F = 0 m2 | Planning cleaning effort | LIBAL database |
Which training room has the largest area? | The training room 00.415 is the largest with an area of 41.473 m2 | Conduction of trainings | LIBAL database |
Which rooms can be used for training? | There are the training rooms 00.415, 01.127 and 01.535 | Conduction of trainings | LIBAL database |
Criterion | Rule for the Criterion Evaluation |
---|---|
Execution Validity | sql_execution_result does not contain “error,” “failed,” “fehler” or “empty value”. |
Correctness | At least one value mentioned from expected_answer appears in the answer OR everything in expected_answer is found in the answer/query. |
Contextual Relevance | Answer shares > 1 meaningful value with expected_answer. |
Hallucination | No meaningful overlap with expected_answer values |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Buga, K.; Tesic, R.; Koyuncu, E.; Hanne, T. Large Language Models for Structured Information Processing in Construction and Facility Management. Electronics 2025, 14, 4106. https://doi.org/10.3390/electronics14204106
Buga K, Tesic R, Koyuncu E, Hanne T. Large Language Models for Structured Information Processing in Construction and Facility Management. Electronics. 2025; 14(20):4106. https://doi.org/10.3390/electronics14204106
Chicago/Turabian StyleBuga, Kyrylo, Ratko Tesic, Elif Koyuncu, and Thomas Hanne. 2025. "Large Language Models for Structured Information Processing in Construction and Facility Management" Electronics 14, no. 20: 4106. https://doi.org/10.3390/electronics14204106
APA StyleBuga, K., Tesic, R., Koyuncu, E., & Hanne, T. (2025). Large Language Models for Structured Information Processing in Construction and Facility Management. Electronics, 14(20), 4106. https://doi.org/10.3390/electronics14204106