Next Article in Journal
Comparison and Interpretability Analysis of Deep Learning Models for Classifying the Manufacturing Process of Pigments Used in Cultural Heritage Conservation
Previous Article in Journal
An Overview of Satellite Link Budget Sensitivity Based on Digital Modulation Schemes in Multi-Orbit Satellite Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Research on Investment Estimation of Prefabricated Buildings Based on Genetic Algorithm Optimization Neural Network

School of Civil Engineering and Architecture, Wuhan Polytechnic University, Wuhan 430023, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(7), 3474; https://doi.org/10.3390/app15073474
Submission received: 8 January 2025 / Revised: 23 February 2025 / Accepted: 11 March 2025 / Published: 21 March 2025

Abstract

At present, China’s prefabricated buildings have entered a period of comprehensive development. Starting from the investment decision-making stage of construction projects, this paper analyses the characteristics of prefabricated investment estimation and the relevant literature on the characteristics of prefabricated construction projects, uses the rough set attribute reduction algorithm to screen the key engineering characteristic factors, and establishes a BP neural network model optimized by genetic algorithm to estimate and analyze the investment of completed prefabricated construction projects. The results of the study show that the accuracy of the improved BP neural network prediction model is better than that of the standard BP neural network prediction model, and the results are more stable, which provides a more scientific and effective method for the investment estimation of prefabricated buildings projects.

1. Introduction

Investment estimation is one of the important contents of the preliminary work of engineering construction, and is an important basis for the feasibility study and investment decision-making [1]. At this stage of China’s prefabricated buildings standardization, integration degree is insufficient, thus resulting in difficult to reduce the construction cost greatly restricts its benign development. In order to further promote the development of prefabricated construction projects and meet the reinvestment estimation needs of prefabricated construction projects, China’s Ministry of Housing and Urban–Rural Development issued the “Investment Estimation Indicators for Prefabricated Construction Projects” in 2023. Therefore, how to accurately control the investment estimation of prefabricated building projects is gradually becoming the focus of research on prefabricated building management. Due to the different production and construction methods between prefabricated buildings and traditional buildings, using traditional methods such as capital turnover rate method, production capacity index method, and comprehensive index investment estimation method to estimate prefabricated buildings is inevitably unreasonable, and digital technology must be introduced for fast and accurate estimation.
In the era of coordinated development of intelligent construction and building industrialization, the innovation and iteration speed of digital technology in China has significantly accelerated. Digital technology is also integrating into various fields of our lives with new concepts, new formats, and new models, and artificial intelligence is one of them. With the rise in artificial intelligence, machine learning models and deep learning techniques have gradually attracted the attention of scholars. As deep learning usually requires a large amount of data for training to fully utilize its advantages, this article selects machine learning for regression prediction of prefabricated building investment estimation. Meanwhile, considering the nonlinear relationship between the characteristic parameters and output indicators of the selected project in this study, a nonlinear model will be selected for prediction in this paper.
Nonlinear models in machine learning typically include methods such as XGBoost, decision trees, random forests, support vector machines, and neural networks. Among them, neural networks, as an important direction of artificial intelligence, have been introduced into the field of engineering investment estimation by many scholars. Based on previous research, this paper will use genetic algorithms to optimize the BP neural network method for investment estimation of prefabricated buildings, solving the limitations of decision tree instability, large memory consumption of random forests, and poor interpretability of support vector machines. At the same time, it enhances global search ability, avoids local optima, avoids overfitting, and improves model training efficiency. In the current era of machine learning, scholars have also begun to use this method to study investment estimation in the engineering field. In addition, scholars’ research on genetic algorithms is constantly deepening.
Yan, HY, and others proposed a prefabricated concrete building investment estimation model based on the XGBoost machine learning algorithm. The construction feature indicators of prefabricated building investment estimation were extracted using the construction project cost significance theory and the analytic hierarchy process. Then, the XGBoost machine learning algorithm prefabricated building investment estimation model was constructed to quantify confidence and prediction uncertainty. Compared with traditional machine learning methods, such as support vector machines, backpropagation neural networks, and random forests, it was found that XGBoost has better generalization and interpretability [2]. Liu, MengKai, and Luo, Meng, used VensimPLE software to establish a system dynamics model. Based on the identification of cost factors in the research review, the ABC classification method and chromatographic analysis were used to score and determine the weights of influencing factors. Finally, it was found that the total cost error obtained by the model was less than 2%, and corresponding strategies were proposed based on the results [3]. Choi, Younguk, Chan Young PARK and Changjun LEE et al. developed a conceptual cost estimation framework consisting of two methods for industrial modular projects by converting strip project information to solve the problem of estimating benefits and costs before making decisions for prefabricated construction projects. They conducted case studies to demonstrate the applicability and accuracy of the framework [4]. Stefan Sievers, Tim Seifert and Marcel Franzen et al. introduced a comprehensive method that combines traditional factor method and modular adjustment to address the limitations of existing estimation methods. Each cost item is decomposed into material, construction, and engineering costs, and the cost adjustment factors brought by modularization are applied. Then, compared with traditional factories and a modular factory containing three production lines, it was found that the total investment cost of a modular factory is 12% higher than that of a traditional factory, but the construction and engineering costs are significantly reduced [5]. Peng Zhang and Yanli Dong have developed a big data supply chain system based on an improved genetic algorithm. Through data preprocessing, multi-objective modeling, and parameter optimization, the system improves the operational efficiency and environmental performance of the supply chain. It has also been found that the algorithm can effectively coordinate resource allocation and reduce environmental impact in green procurement, ecological design, and other links [6]. Yunlong Wang et al. proposed a PSOIG method that combines particle swarm optimization and genetic algorithm to solve three-dimensional ship pipeline layout problems. By constructing cabin space models, obstacle attitude space models, and directional guidance mechanisms, the efficiency of path search is optimized, and cross and mutation strategies are introduced to enhance global search capabilities. The results show that compared with traditional ant colony algorithms, PSOIG performs better in path length, number of bends, and obstacle avoidance performance, with a convergence improvement speed of over 24%. It can efficiently generate pipeline layout schemes that meet engineering constraints, providing new ideas for intelligent ship design [7]. Baogang Wang et al. proposed an improved NSGA-II algorithm to address the problems of weak convergence ability and poor single objective optimization effect of traditional genetic algorithms in solid wood panel layout optimization. The algorithm generates a reverse population through reverse learning to enhance search ability and combines directional mutation and uniform mutation strategies to improve population diversity and optimization efficiency. The results show that the improved algorithm can simultaneously optimize multiple objectives such as material utilization, economic value, and size priority while reducing the number of convergence iterations, significantly improving wood utilization and enterprise efficiency [8]. Bosong Duan et al. proposed a new parallel hybrid genetic particle swarm optimization algorithm for solving multi constraint optimization problems. By combining the efficiency of PSO and the global optimization capability of GA through a parallel architecture, PSO was used in the early stage of optimization, and GA was switched to update particles when trapped in local optima. The results showed that this algorithm has significant advantages in finding optimal values, convergence speed, and time overhead when solving multi constraint optimization problems [9]. Sethembiso Nonjabulo Langazane and Akshay Kumar Saha proposed a comprehensive sensitivity analysis to address the problem of premature convergence caused by improper selection of control parameters in solving the coordination problem of overcurrent relays using particle swarm optimization and genetic algorithms. The analysis evaluated the impact of discrete control parameters on the performance of particle swarm optimization and genetic algorithms as well as their effects on the behavior of overcurrent relays. The results showed that particle swarm optimization algorithm is more sensitive to inertia weight and population size, and genetic algorithm produces faster convergence speed in the case of 30% crossover, 2% mutation, and small population size. The performance of genetic algorithm was improved by optimizing the fitness function, and the sensitivity of particle swarm optimization parameter settings was verified to be better than that of genetic algorithm [10]. Jaecheon Kim et al. proposed an improved real coding genetic algorithm to solve bridge model optimization problems by developing a new dynamic mutation operator. The algorithm combines crossover and mutation operators, dynamically adjusting the mutation mode during the optimization process to improve the convergence speed and search efficiency of the global optimal solution. The performance of the algorithm and its effectiveness in bridge model optimization were verified through various testing problems [11]. Gwang Hee Kima et al. applied a backpropagation neural network model based on genetic algorithm to cost estimation. After training and evaluating the performance of the model using actual engineering data, it was shown that the BPN model using GA was more effective and accurate in building cost estimation than the BPN model using trial and error method [12]. Zhou Chengjie explored an engineering investment estimation method based on the BP neural network model, which has a highly nonlinear relationship with the main characteristic factors of the project. By verifying the quantification and prediction of statistical data of existing railway bridge projects, it was shown that the model meets the requirements for rapid estimation of engineering investment [13]. Pan Yuhong et al. introduced the GA–BP model into the cost estimation of highway engineering to address the existing problems. They verified through examples that the prediction accuracy of the GA–BP algorithm is higher than that of the BP neural network, and it is feasible and effective [14]. Zhang Feilian and Liang Xiufeng proposed a genetic algorithm (GA) to optimize the input weights and hidden layer thresholds of extreme learning machines (ELMs). Through training and validation of urban rail transit engineering sample data, it was shown that the accuracy and stability of this estimation model were significantly improved compared to other estimation models [15]. Wang Fuyu et al. proposed a BP neural network model based on the Tianniu Xu search algorithm. After conducting investment estimation analysis on typical assembly cases, it was shown that the BAS-BP network model had higher prediction accuracy than the BP neural network model, and the results were more stable [16].
From the above literature analysis, it can be seen that the current research on investment estimation of prefabricated buildings is a combination of quantitative methods and machine learning or system dynamics methods, and most of the research on BP neural networks is focused on transportation rail and traditional buildings. Meanwhile, genetic algorithms also demonstrate their performance advantages by improving or combining them with other optimization methods. Therefore, based on the profound theoretical foundation of BP neural network and genetic algorithm, this article proposes a genetic algorithm-optimized BP neural network for prefabricated building investment estimation prediction model. By training and testing data samples of prefabricated building engineering, the feasibility of the model is proved, providing reference for accurate prediction of investment estimation of prefabricated buildings.

2. Improved BP Neural Network Model Based on GA

2.1. BP Neural Network Principle

BP neural network is a multi-layer feedforward neural network trained by error backpropagation algorithm, which repeatedly adjusts the weights and thresholds of the network through backpropagation process to minimize the difference between the output value and the true value [17]. The topology structure of the BP neural network model includes an input layer, a hidden layer, and an output layer. The hidden layer can have one or more layers, and the layers are fully interconnected. The basic structure is shown in Figure 1 [18]. In the figure, two types of signals propagate between layers, namely signal forward propagation and error backpropagation, based on the difference between the input and weight functions and the actual output value and expected value of the network. In the forward propagation process, the input vector starts from the hidden layer, is calculated layer by layer through the hidden layer, and then transmitted to the output layer, and the state of each layer of neurons only affects the state of the next layer of neurons. When the output layer cannot obtain the expected value, it will enter the error backpropagation process, and the error signal will return along the original route, gradually adjusting the weights and thresholds of each layer of the network until the input layer and then repeating iterative calculations.
The specific operation steps are as follows.
1. Select training samples and test samples to provide to the network;
2. Positive propagation:
(1) The input neti of the i-th node in the hidden layer:
n e t i = j = 1 M ω i j x j + θ i .
(2) The output Yi of the i-th node in the hidden layer:
Y i = Φ n e t i = Φ j = 1 M ω i j x j + θ i .
The commonly used activation functions include Sigmoid function and ReLU function.
(3) The input netk of the kth node in the output layer is
n e t k = i = 1 q ω k i y i + a k = i = 1 q ω k i Φ j = 1 M ω i j x j + θ i + a k .
(4) The output of the kth node in the output layer, Yk, is
Y k = ψ n e t k = ψ i = 1 q ω k i Φ j = 1 M ω i j x j + θ i + a k .
3. Backpropagation:
(1) The error function of the output layer is defined as
E = 1 2 k = 1 L z k y k 2 .
(2) The total error function of the system for the training samples is
E = 1 2 p = 1 p k = 1 L z k y k 2 .
(3) Correct the ownership values and thresholds obtained from the output layer and hidden layer:
Δ ω k i = η E ω k i = η E n e t k n e t k w k i = η E y k y k n e t k n e t k w k i
Δ a k = η E a k = η E n e t k n e t k a k = η E y k y k n e t k n e t k a k
Δ ω i j = η E ω i j = η E n e t i n e t i ω i j = η E y i y i n e t i n e t i ω i j
Δ θ i = η E θ i = η E n e t i n e t i θ i = η E y i y i n e t i n e t i θ i .
The organized formula is
Δ ω k i = η p = 1 p k = 1 L z k p y k p ψ n e t k y i
Δ a k = η p = 1 p k = 1 L z k p y k p ψ n e t k
Δ ω i j = η p = 1 p k = 1 L z k p y k p ψ n e t k ω k i Φ n e t i x i
Δ θ i = η p = 1 p k = 1 L z k p y k p ψ n e t k ω k i Φ n e t i .
The BP neural network algorithm is widely used because of its simple operation and strong parallelism, but because it chooses the fastest descent method in nonlinear programming and modifies the weight according to the negative gradient direction of the error function, it has problems such as slow convergence speed and easy to fall into a local minima state. Therefore, this article uses the GA to optimize its weights and thresholds to solve the above problems and improve the prediction accuracy of the network model.

2.2. Principles of Genetic Algorithm

Genetic algorithm imitates the evolutionary mechanisms of biological selection, crossover, and mutation to find the most suitable individuals. It is a random search algorithm that can solve optimization problems for both discrete and continuous functions. In genetic algorithms, the solution to each problem is called a string of chromosomes, typically represented as a bit string, and each chromosome is referred to as an individual. This algorithm adopts a population organization search method, where these populations (chromosome strings) are composed of several individuals, and the individuals are composed of units that control one or more genetic features. The genes for certain features are distributed along the chromosome, and the corresponding strings are called loci. Each genotype represents a potential solution to a problem, which is obtained by mapping chromosome representations to the decision variable space through fitness functions [19]. Algorithms usually preset four operating parameters: population size, evolutionary generation, crossover probability, and mutation probability. When the preset parameter limits are reached, the algorithm will automatically stop. The basic process of classical genetic algorithms is shown in Figure 2.
The specific operation steps are as follows.
(1) Genetic Algorithm Encoding
In genetic algorithm optimization of BP neural network algorithm, each chromosome is decomposed into two parts: connecting genes and parameter genes, and different encoding methods are adopted for these two parts. Encoding is a method of transforming feasible solutions of a problem from its solution space to the search space that the algorithm can handle using genetic algorithms. It is the first problem to be solved when using genetic algorithms. Meanwhile, encoding also affects the operations of selection, crossover, and mutation operators during the operation of genetic algorithms.
(2) Design of fitness function
The fitness function, also known as the evaluation function, in genetic algorithms is the only method to determine the quality of solutions by judging the degree of superiority or inferiority of individuals in the population based on the objective function of the problem being solved. It has a direct impact on the convergence speed of the algorithm and the search for the optimal solution. In the process of searching evolution, genetic algorithms use fitness function values to measure the degree to which each individual in the population can achieve or approach finding the optimal solution in optimization calculations. The fitness function in this article is based on the total error of the neural network and is expressed as follows:
f = 1 1 + E .
Among them, E is the total error in the network structure, Z j k is the ideal output, Y j k is the actual output, and K is the number of sample sets.
(3) Genetic operator design
Genetic operations include three basic genetic operators: selection, crossover, and mutation.
(1) The operation of finding excellent individuals from a group and eliminating inferior individuals is called selection. The purpose of selection is to directly pass on optimized individuals or solutions to the next generation or generate new individuals through paired crossover and then pass them on to the next generation. This operation is based on the assessment of individual fitness in the population. The selection operator in this article adopts the most commonly used and simplest roulette wheel selection in genetic algorithms.
The roulette wheel selection first calculates the selection probability of each individual, and the steps are as follows.
① Calculate the sum of fitness values for all individuals in the population:
F = K = 1 M E v a l ν k , k = 1 , 2 , , M .
Among them,
F is the sum of the fitness values of all individuals in the group;
M is the number of individuals in the group;
E v a l ν k is the fitness value of the kth individual in the population.
② Calculate the selection probability for each chromosome:
p k = E v a l v k F .
Among them, pk is the selection probability of the kth chromosome.
After calculating the selection probability of each chromosome, construct a disk and divide it into N sectors, with the center angle of the Kth sector being 2π p k . Then, imagine a pointer pointing to the junction of the first sector and the Nth sector. Rotate the disk counterclockwise M times, with each rotation. When the disk stops rotating, if the pointer points to the kth sector, select the kth chromosome. When the pointer points to the boundary between the kth sector and the k+1st sector, the kth chromosome is selected. In the algorithm implementation of roulette wheel selection, a random number r from [0, 1] is used to simulate the position pointed by the pointer after the disk stops rotating, and the sector pointed by the pointer is determined by calculating the cumulative probability of each chromosome ν k .
(2) The crossover operator plays a core role in genetic algorithms, which replaces and recombines partial structures of two parent individuals to generate new individuals. Through this operation, the search ability of genetic algorithms can be greatly improved. The commonly used crossover operators include single-point crossover, two-point crossover, multi-point crossover, and uniform crossover. In this article, single-point crossover is used. The so-called single-point crossover refers to randomly selecting a crossover point among individuals and then swapping the substrings of two parents on the right side of the hybridization point to generate two offspring.
In genetic algorithm, single-point crossover first selects several chromosomes from the population generated by roulette wheel selection with a certain probability as the parent, and then generates M random numbers from [0, 1]. If the probability of selecting the parent is greater than or equal to the generated random number, the chromosome corresponding to the random number in the population is selected as the parent. If the selected parents are an odd number, then randomly select another parent from the group, or delete one from the already selected parents. Thus, select an even number of parents and cross them pairwise.
(3) The basic content of mutation operators is to change the gene values on certain loci of individual strings in a population with the aim of introducing chromosome diversity in the population and preventing immature convergence of the algorithm. According to different encoding representations of individuals, it can be divided into real valued variation and binary variation. The operation process of the mutation operator in genetic algorithm is to generate a random number r for each gene locus of each chromosome in the population after hybridization. If r p m , the gene locus is mutated, otherwise no mutation is performed.
Before the evolution of genetic algorithms, the solution of the solution space was represented as genotype string structure data of genetic algorithms, and different combinations of these string structure data constituted different loci. After setting the 4 operating parameters, the algorithm will randomly generate N individuals as the initialized population, reflect the advantages and disadvantages of the individual or solution through the appropriate fitness function, and then use three genetic operators, namely selection, crossover and mutation to form a new population with excellent individuals, and then return to the fitness assessment step. When the convergence criterion is reached, the algorithm automatically stops iteration and terminates the operation.

2.3. Construction of the GA–BP Neural Network Model

As a random search algorithm, the genetic algorithm has a strong global merit ability and robustness, which can optimize the weights and thresholds of BP neural networks, which are mainly divided into three parts: the determination of the structure of the BP neural network, the optimization of weights and thresholds of genetic Algorithms, and the training and prediction of BP neural networks. In this paper, the genetic algorithm is used to optimize the initial weights and thresholds of the BP neural network. The optimal weights and thresholds are obtained, and the optimal weights and thresholds are then applied to the BP neural network for training, which can prevent the network from falling into local optimization and improve the training performance. For most problems, the BP neural network with a three-layer structure is sufficient, so a three-layer neural network is selected in this paper, including the input layer, the hidden layer, and the output layer. In the following example, the number of engineering characteristic factors of property reduction in the vector of the rough set is the input layer, per square meter cost of the project is the output layer node, and the hidden layer node N is determined according to the empirical formula, where N is the number of hidden layer nodes, m is the number of neuron nodes of the input layer, and n is the number of output layer neuron nodes, which is N = m + n + α , α , a constant between them [1, 10]. Based on the above theory, the steps of genetic algorithm to optimize BP neural network are as follows.
(1) After determining the topology of the neural network, initial weights and thresholds are generated. They are expressed in the BP neural network in the form of vectors, including the vectors composed of the connection weights W1j,…, Wij between the input layer and the hidden layer, the vectors composed of hidden layer threshold θ1, …, θi, the vectors composed of the connection weights W2j, …, Wkj between the hidden layer and the output layer, and the vectors composed of the threshold of the output layer a1, …, ak.
(2) Determine the fitness function. The selection of the fitness function is determined by the minimum mean square error between the obtained test set and the training set by the BP neural network on the data, and the smaller the error, the higher the accuracy.
(3) Determine the global search operator of the algorithm. The roulette method is used to select individuals with good fitness from the population. Two chromosomes are randomly selected from the population, a cross point is randomly generated at the same time, the other half of the two chromosomes is exchanged to obtain the two newly generated individuals, and their feasibility is tested. For the mutation operation, one individual is randomly selected and a mutation point is randomly selected, and the mutation is carried out according to a certain probability, so as to produce a new individual [20].
(4) If the obtained neural network weights and thresholds do not meet the convergence conditions, return to step (1) and iterate again.
(5) After obtaining the optimal weights and thresholds of the BP neural network, they are assigned to the BP neural network for optimized training and prediction to obtain the final results.
Figure 3 shows the process.

3. Determination of Engineering Eigenvectors

It can be seen from the “Investment Estimation Index of Prefabricated Construction Engineering” that the index base price of investment estimation of prefabricated concrete structure engineering is composed of construction engineering cost, installation engineering cost, equipment purchase cost, other engineering construction costs, and basic preparation costs. Also, the proportion of construction and installation engineering cost to the index base price is over 80%. The selection of engineering features directly affects the prediction accuracy of the model, which should reflect not only the basic characteristics of the project, but also the core parameters that affect the cost of the project. According to the research of Hu Guoqing et al. [21], it shows that the cost-significant items (CSIs) of prefabricated buildings are reflected primarily in four projects: cast-in-place reinforced concrete engineering, PC component and installation, decoration engineering, and water supply and drainage engineering., The cost of these four CSIs represents more than 85% of the total cost. Therefore, based on analyzing the cost composition of traditional buildings and prefabricated building projects, this article preliminarily screens out 11 characteristic parameters that affect the estimation of investment estimation of prefabricated buildings by reviewing the literature [16,22,23]: foundation type, structure type, construction area, number of building floors, prefabricated component type, prefabricated component connection method, roofing and waterproofing, decoration engineering, installation, electrical and water supply and drainage engineering, assembly rate, and project cost index.
Rough set theory is the work of the Polish scholar Z. Pawlak, proposed in 1982. It can deal with the problems of data simplification and inconsistent information analysis on the premise of retaining information. It has been widely used in machine learning, knowledge acquisition, decision analysis, and other fields [22]. To avoid the problem of data redundancy caused by too many factors of engineering features, this paper adopts rough set theory to reduce the dimension of engineering feature indicators and improve the speed of the neural network. Since the attribute values in the decision table of rough set theory need to be discretized, the above qualitative features are quantified in this paper, and the quantitative results are shown in Table 1.
The decision table is an important knowledge expression system in rough set theory. The basic structure of the decision table in this paper is based on engineering characteristic factors as conditional attributes, and the unilateral cost of prefabricated engineering as the decision attribute. By referring to prefabricated construction engineering case studies [24,25] and collecting information on some newly built prefabricated construction projects, 16 sets of prefabricated construction project data from central and southern China were finally compiled for network training, including 15 sets as training samples and 1 set as a testing sample. Due to the confidentiality of some information in the construction project, it is required not to disclose the company and project to which it belongs. Therefore, this article combines the complete actual engineering case information collected offline with the data cited in the literature. At the same time, the Boolean reasoning algorithm in the Rosetta rough set software, adapted to Windows 2000/XP version, is used to discretize the six continuous variables of building area, building floors, installation, electrical and plumbing engineering, assembly rate, engineering cost index, and decision attribute engineering unit cost in the conditional attributes, and it establishes an initial decision table, as shown in Table 2.
Once the initial decision table is established, genetic algorithm reduces its properties, and the results are shown in Table 3 [26]. Then, select the frequency of ≥14 in the table as the main engineering feature index and use it as the input vector of the neural network.

4. Case Studies

4.1. Neural Network Data Preprocessing

Due to the significant differences and inconsistent units in the data of the main engineering feature indicators, the learning speed and prediction accuracy of the neural network are greatly limited. Therefore, the normalization formula is used to preprocess the data in Table 4, mapping the row minimum and maximum values to [−1, 1] to process the input engineering feature vectors. The processing results are shown in Table 5:
X = x m i n x m a x x m i n

4.2. Model Simulation and Comparative Analysis

Using MATLAB R2019a software, 15 sets of sample data were used to optimize the weights and thresholds of the BP neural network through GA genetic algorithm for simulation training, and one set of data was left for prediction verification. The predicted values after fixing the random seeds are shown in Table 6, and the errors between the training set and the test set are shown in Table 7. (Among them, based on empirical formulas, the network prediction model is determined to have 10 hidden layer nodes, and the parameters of GA genetic algorithm are as follows. Population size: 50; maximum iteration number: 30; intersection probability: 0.8; mutation probability: 0.2).
On the basis of optimizing the BP neural network with genetic algorithm, the PSO optimization algorithm was used to predict and analyze the model. The results are shown in Table 8, and the errors between the training set and the test set are shown in Table 9:
Based on the small sample size of this study, stratified k (k = 5) fold cross validation was used, and the optimization results of the two optimization algorithms are shown in Table 10.
According to Table 10, in the case of a small sample size in this study, the absolute value of the MSE of the genetic algorithm is relatively high, with large fluctuations between different folds. A high MSE standard deviation indicates that the model is sensitive to data partitioning, but the MAPE of the model shows good relative error control. On the other hand, the results of PSO are very poor, and the model can hardly make accurate predictions. Based on this, it can be inferred that genetic algorithms can better adapt to small sample problems by searching the solution space through selection, crossover, and mutation. PSO relies on particle swarm intelligence, which may be difficult to converge in large parameter spaces, especially when there are many local optima in high-dimensional spaces.
The above results are all the results of running the model multiple times. During the simulation process, although the relevant parameters of the model were modified multiple times, the cross-validation results were always unsatisfactory. Therefore, it can be concluded that the optimization performance of genetic algorithm is superior to PSO algorithm in this problem.

5. Conclusions

This article combines prefabricated building investment estimation, rough set theory, and the GA–BP network prediction model. Firstly, based on the collection of factors affecting the characteristics of prefabricated construction engineering, this article uses rough set theory to perform attribute reduction, which improves the prediction accuracy of the GA–BP model while eliminating redundant attributes. Secondly, by normalizing the initial sample data through function operations, the GA–BP neural network model is applied to predict the investment estimation of prefabricated buildings, and the results are compared with the PSO-optimized BP neural network model. The results indicate that the overall prediction accuracy of the GA–BP neural network model in this study is better than that of the PSO-optimized BP neural network, and its prediction results are more stable. It is scientifically effective for estimating and predicting prefabricated buildings in the investment decision-making stage.
The complete investment estimation indicators for prefabricated buildings include equipment purchase costs and other construction costs, among others. As the construction and installation costs of prefabricated buildings account for 80% or more of the base price of the investment estimation prediction indicators for prefabricated buildings, this article only predicts the construction and installation costs of prefabricated buildings. On the other hand, the prediction model of neural networks is built on the basis of obtaining complete information from completed prefabricated construction projects. When collecting relevant data on prefabricated construction projects, this article found that the engineering data information provided by relevant construction units was incomplete, and the information on prefabricated construction projects on relevant construction websites and government regulatory platforms was also incomplete. Therefore, this article suggests that, in actual engineering construction, all participating units should timely exchange engineering information, achieve transparency of engineering information in accordance with principles, and upload it to government regulatory platforms in a timely manner. At the same time, the government should strengthen the supervision of prefabricated construction projects, ensure the timeliness and effectiveness of information uploaded by relevant units participating in prefabricated construction, and improve the database of prefabricated construction. In addition, a BIM-based data platform for the entire lifecycle of prefabricated buildings can be established, integrating data flows from design, production, logistics, and assembly to achieve dynamic mapping of identification codes and cost information.
However, this article still has certain limitations. Firstly, there is the issue of sample size, as a small number of samples may lack certain persuasiveness. Secondly, there is the research methodology. This article uses the genetic optimization algorithm. In future research, if there are sufficient data, deep-learning methods can be introduced for prediction.

Author Contributions

Writing—original draft preparation, J.G.; writing—review and editing, W.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The manuscript has indicated the source of the data.

Acknowledgments

Special thanks to Wang Lei of Wuhan Construction Engineering Emerging Building Materials Green Industry Technology Co., Ltd. for providing practical engineering case experience for this article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhang, M.; Liang, J. Investment Estimation Method of Construction Project in Decision Stage. Pet. Plan. Eng. 2005, 53–55. [Google Scholar]
  2. Yan, H.; He, Z.; Gao, C.; Xie, M.; Sheng, H.; Chen, H. Investment estimation of prefabricated concrete buildings based on XGBoost machine learning algorithm. Adv. Eng. Inform. 2022, 54, 101789. [Google Scholar]
  3. Liu, M.; Luo, M. Cost estimation model of prefabricated construction for general contractors based on system dynamics. Eng. Constr. Archit. Manag. 2025, 32, 621–638. [Google Scholar]
  4. Choi, Y.; Park, C.Y.; Lee, C.; Yun, S.; Han, S.H. Conceptual cost estimation framework for modular projects: A case study on petrochemical plant construction. J. Civ. Eng. Manag. 2022, 28, 150–165. [Google Scholar]
  5. Sievers, S.; Seifert, T.; Franzen, M.; Schembecker, G.; Bramsiepe, C. Fixed capital investment estimation for modular production plants. Chem. Eng. Sci. 2017, 158, 395–410. [Google Scholar]
  6. Zhang, P.; Dong, Y. Strategy transformation of big data green supply chain by using improved genetic optimization algorithm. Soft Comput. 2023, 1–10. [Google Scholar] [CrossRef]
  7. Yunlong, W.; Hao, W.; Guan, G.; Li, K.; Lin, Y.; Chai, S. Intelligent Layout Design of Ship Pipeline Using a Particle Swarm Optimisation Integrated Genetic Algorithm. Int. J. Marit. Eng. 2021, 163. [Google Scholar] [CrossRef]
  8. Wang, B.; Yang, C.; Ding, Y. Non-Dominated Sorted Genetic Algorithm-II Algorithm-based Multi-Objective Layout Optimization of Solid Wood Panels. BioResources 2022, 17, 94–108. [Google Scholar]
  9. Duan, B.; Guo, C.; Liu, H. A hybrid genetic-particle swarm optimization algorithm for multi-constraint optimization problems. Soft Comput. 2022, 26, 11695–11711. [Google Scholar]
  10. Langazane, S.N.; Saha, A.K. Effects of Particle Swarm Optimization and Genetic Algorithm Control Parameters on Overcurrent Relay Selectivity and Speed. IEEE Access 2022, 10, 4550–4567. [Google Scholar]
  11. Deep, K.; Thakur, M. Development of a Mutation Operator in a Real- Coded Genetic Algorithm for Bridge Model Optimization. Appl. Math. Comput. 2007, 193, 211–230. [Google Scholar]
  12. Kim, G.H.; Yoon, J.E.; An, S.H.; Cho, H.H.; Kang, K.I. Neural network model incorporating a Genetic Algorithm in estimating construction costs. Build. Environ. 2004, 39, 1333–1340. [Google Scholar]
  13. Zhou, C. Discussion of Project Investment Estimation Method based on BP Neural Network. Railw. Eng. Cost Manag. 2015, 30, 6–9+13. [Google Scholar]
  14. Pan, Y.H.; Zhang, Y.L.; Cai, Y.J.; Wu, H.H.; Sui, H.Y. Research on Highway Engineering Cost Estimation Based on the GA-BP Algorithm. J. Chongqing Jiaotong Univ. (Nat. Sci.) 2016, 35, 141–145. [Google Scholar]
  15. Zhang, F.; Liang, X. Research on investment estimation method of urban rail transit project based on GA-ELM. J. Railw. Sci. Eng. 2019, 16, 1842–1848. [Google Scholar]
  16. Wang, F.; Zhang, S.; Li, Y. Application of Neural Network Model of BAS-BP in Investment Estimation of Prefabricated Building. J. Anhui Univ. Technol. (Nat. Sci.) 2019, 36, 382–387. [Google Scholar]
  17. Rumelhart, D.; Hinton, G.; Williams, R. Learning representations by back-propagating errors. Nature 1986, 323, 533–536. [Google Scholar]
  18. MATLAB Technology Alliance; Liu, B.; Guo, H. MATLAB Neural Network Super Learning Manual; Posts & Telecom Press: Beijing, China, 2014; pp. 159–161. [Google Scholar]
  19. Scrucca, L. GA: A package for Genetic Algorithms in R. J. Stat. Softw. 2013, 53, 1–37. [Google Scholar]
  20. Xiao, Y.; Yang, J.; Zhou, S. Engineering Optimization—Theory, Model and Algorithm; Beihang University Press: Beijing, China, 2021; Volume 1, p. 317. [Google Scholar]
  21. Hu, Q.; Cai, B.; He, Z.; Dai, Z. Investment estimation of prefabricated building based on BP neural network. J. Chang. Univ. Sci. Technol. (Nat. Sci.) 2018, 15, 66–72+86. [Google Scholar]
  22. Han, Z.; Zhang, Q.; Wen, F. Rough sets: Theory and application. Inf. Control 1998, 38–46. [Google Scholar]
  23. Hu, Q.; Tian, X.; He, Z. Green building investment estimation method based on genetic algorithm optimized extreme learning machine. Build. Econ. 2020, 41, 125–130. [Google Scholar]
  24. Xia, F.; Fang, S.; Chen, P. Typical Project Case About Prefabricated Building. Hous. Sci. 2015, 35, 18–23. [Google Scholar]
  25. Zhao, W.; Wang, S. Research on the Investment Estimation of Prefabricated Building Based on BASBP Model. J. Anhui Univ. Sci. Technol. (Nat. Sci.) 2020, 40, 73–79. [Google Scholar]
  26. Xu, W.; Shen, J.; Chen, X.; Wan, S. Construction cost prediction model based on RS-RBFNN. J. Qingdao Univ. Technol. 2021, 42, 96–102. [Google Scholar]
Figure 1. Structure diagram of a typical three-layer BP neural network.
Figure 1. Structure diagram of a typical three-layer BP neural network.
Applsci 15 03474 g001
Figure 2. Genetic Algorithm flow chart.
Figure 2. Genetic Algorithm flow chart.
Applsci 15 03474 g002
Figure 3. Flow diagram of the GA–BP neural network.
Figure 3. Flow diagram of the GA–BP neural network.
Applsci 15 03474 g003
Table 1. Quantification of characteristics of prefabricated building projects.
Table 1. Quantification of characteristics of prefabricated building projects.
Engineering FeaturesQuantify the Value
12345
Foundation typeIndependent foundationRaft foundationStrip foundationsPile foundations
Structure typeShear wall structureFrame structureFrame shear wall structureTube structurePrefabricated containerized structure
AreaEnter based on actual data, unit: m2
Number of floors of the buildingEnter based on actual data
The Type of prefabricated componentPrefabricated laminated floor slabsPrefabricated sandwich insulated exterior wall panelsPrefabricated interior wall panelsPrefabricated stairsPrefabricated air conditioning panels
Prefabricated component connectionsGrouting rebar sleeve connectionReinforcement slurry anchor lap connection
Roofing and waterproofingFlat roof
Class 1 waterproof
Flat roof
Class 2 waterproof
Flat roof
Class 3 waterproof
Pitched roof
Class 1 waterproof
Curved roofs
Class 1 waterproof
Decoration worksSimple decorationStandard finishMid-range finishHigh-end decoration
Installation, Electrical and Water Supply and Drainage Engineering/(Yuan·m−2).Enter based on actual data
Assembly rate/%.Enter based on actual data
Project cost indexEnter based on actual data
Table 2. Initial decision table.
Table 2. Initial decision table.
Sample No.Conditional AttributesDecision
Attributes
Foun-dation
Type
Struc-ture TypeAreaNumber of Floors of the BuildingThe Type of Prefabric-ated ComponentPrefabricated Component ConnectionsRoofing and WaterproofingDecoration WorksInstallation, Electrical and Water Supply and Drainage Works (RMB/m2).Asse-mbly rate (%)Project Cost IndexCost per Square Meter (RMB/m2)
141121 + 4 + 51113123
211341 + 31121431
321311 + 2 + 3 + 41123235
421121 + 2 + 3 + 42121124
524311 + 2 + 31113132
611341 + 32121421
721121 + 2 + 31121123
823531 + 41215123
923321 + 41215425
1041211 + 4 + 51111122
1124421 + 2 + 31121131
1221111 + 2 + 31421132
1311541 + 2 + 32421132
1421211 + 2 + 31121131
1521421 + 31123432
1621321 + 31111232
Table 3. Brief table of initial engineering characteristics.
Table 3. Brief table of initial engineering characteristics.
Initial FeaturesFrequencyInitial FeaturesFrequency
Base type8Structure type9
Area21Number of floors of the building16
The Type of prefabricated component16Prefabricated component connections21
Roofing and waterproofing17Decoration works14
Installation, electrical and plumbing works12Assembly rate14
Project cost index17
Table 4. Data table of sample indicators of main engineering characteristics.
Table 4. Data table of sample indicators of main engineering characteristics.
Sample NumberAreaNumber of Floors of the BuildingThe Type of Prefabricated ComponentPrefabricated Component ConnectionsRoofing and WaterproofingDecoration WorksAssembly Rate (%)Project Cost IndexCost per Square Meter (RMB/m2)
121,749251 + 4 + 5111511002033.03
213,841181 + 3112501052079.71
322,164331 + 2 + 3 + 4112651092286.57
420,020301 + 2 + 3 + 4212681132265.35
523,860331 + 2 + 3111631052104.47
614,638181 + 3212501131997.14
719,908301 + 2 + 3112621132195.35
8316761 + 4121301002805
915,733271 + 4121321002743
1020,453341 + 4 + 5111511002192.68
1116,451271 + 2 + 3112601052094.53
1220,246331 + 2 + 3142601092153.62
1312,881181 + 2 + 3242601052135.28
1421,468331 + 2 + 3112581092076.81
1518,200271 + 3112501052006.45
1619,309301 + 3111551092073.73
Table 5. Normalization table of sample data of main engineering characteristic indicators.
Table 5. Normalization table of sample data of main engineering characteristic indicators.
Sample NumberAreaNumber of Floors of the BuildingThe Type of Prefabricated ComponentPrefabricated Component ConnectionsRoofing and WaterproofingDecoration WorksAssembly Rate (%)Project Cost IndexCost per Square Meter (RMB/m2)
10.7960.3571.000−1.000−1.000−1.0000.105−1.000−0.911
20.032−0.143−1.000−1.000−1.0001.0000.053−0.231−0.796
30.8360.9291.000−1.000−1.0001.0000.8420.385−0.283
40.6290.7141.0001.000−1.0001.0001.0001.000−0.336
51.0000.929−0.333−1.000−1.000−1.0000.737−0.231−0.734
60.109−0.143−1.0001.000−1.0001.0000.0531.000−1.000
70.6180.714−0.333−1.000−1.0001.0000.6841.000−0.509
8−1.000−1.000−0.667−1.000−0.333−1.000−1.000−1.0001.000
90.2150.500−0.667−1.000−0.333−1.000−0.895−1.0000.847
100.5600.714−1.000−1.000−1.000−1.0000.3160.385−0.810
110.2840.500−0.333−1.000−1.0001.0000.579−0.231−0.759
120.6510.929−0.333−1.0001.0001.0000.5790.385−0.613
13−0.061−0.143−0.3331.0001.0001.0000.579−0.231−0.658
140.7690.929−0.333−1.000−1.0001.0000.4740.385−0.803
150.4530.500−1.000−1.000−1.0001.0000.053−0.231−0.977
160.6711.0001.000−1.000−1.000−1.0000.105−1.000−0.516
Table 6. GA optimized BP neural network prediction results.
Table 6. GA optimized BP neural network prediction results.
NumberReal ValuePredicted ValueNumberReal ValuePredicted Value
12033.032092.03927432624.79
22079.712052.06102192.682137.67
32286.572280.84112094.532031.71
42265.352212.65122153.622130.39
52104.471997.63132135.282138.09
61997.142196.37142076.812149.30
72195.352162.74152006.452034.42
828052637.37Test set2073.732053.40
Table 7. GA optimized BP neural network simulation error table.
Table 7. GA optimized BP neural network simulation error table.
NameMSEMAEMAPE (%)
Training set7657.2767.603.02
Test set413.4920.330.98
Table 8. PSO optimized BP neural network prediction results.
Table 8. PSO optimized BP neural network prediction results.
NumberReal ValuePredicted ValueNumberReal ValuePredicted Value
12033.032453.59927431947.21
22079.713923.95102192.681092.39
32286.573552.40112094.533542.95
42265.358047.56122153.626431.30
52104.473538.07132135.283401.63
61997.141471.92142076.813542.92
72195.353269.72152006.453544.70
82805770.62Test set2073.732010.36
Table 9. PSO optimized BP neural network simulation error table.
Table 9. PSO optimized BP neural network simulation error table.
NameMSEMAEMAPE (%)
Training set4,973,203.321751.5579.29
Test set4015.4363.373.06
Table 10. Cross-validation results.
Table 10. Cross-validation results.
Optimization AlgorithmMSEMAEMAPE
genetic algorithm9528.55 ± 8926.3173.35 ± 44.563.3 ± 2.0%
PSO algorithm1,406,319.31 ± 626,556.231123.71 ± 245.0651.1 ± 9.6%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gao, J.; Zhao, W. Research on Investment Estimation of Prefabricated Buildings Based on Genetic Algorithm Optimization Neural Network. Appl. Sci. 2025, 15, 3474. https://doi.org/10.3390/app15073474

AMA Style

Gao J, Zhao W. Research on Investment Estimation of Prefabricated Buildings Based on Genetic Algorithm Optimization Neural Network. Applied Sciences. 2025; 15(7):3474. https://doi.org/10.3390/app15073474

Chicago/Turabian Style

Gao, Jin, and Wanhua Zhao. 2025. "Research on Investment Estimation of Prefabricated Buildings Based on Genetic Algorithm Optimization Neural Network" Applied Sciences 15, no. 7: 3474. https://doi.org/10.3390/app15073474

APA Style

Gao, J., & Zhao, W. (2025). Research on Investment Estimation of Prefabricated Buildings Based on Genetic Algorithm Optimization Neural Network. Applied Sciences, 15(7), 3474. https://doi.org/10.3390/app15073474

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop