Next Article in Journal
Comparative Bibliometric Analysis of Biomethane Production from Anaerobic Digestion of Pig Slurry and Slaughterhouse Wastewater: Research Trends and Gaps
Previous Article in Journal
A Lightweight CNN Pipeline for Soil–Vegetation Classification from Sentinel-2: A Methodological Study over Dolj County, Romania
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hybrid GA-BP Neural Network for Accurate Prediction of TBM Advance Speed Under Complex Geological Conditions

1
Power China Chengdu Engineering Corporation Limited, Chengdu 610072, China
2
College of Water Resource and Hydropower, Sichuan University, Chengdu 610065, China
3
State Key Laboratory of Hydraulics and Mountain River Engineering, Sichuan University, Chengdu 610065, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(22), 12115; https://doi.org/10.3390/app152212115
Submission received: 14 October 2025 / Revised: 7 November 2025 / Accepted: 9 November 2025 / Published: 14 November 2025

Abstract

TBM construction projects require substantial investment, making the accurate and rational prediction of TBM advance speed essential for cost control and timely project completion. To address this, the study develops a precise predictive model for TBM advance speed by integrating the BP neural network model with the genetic algorithm. Initially, raw TBM data were processed to remove non-operational records and anomalous values recorded during construction, resulting in a refined database of TBM operational parameters. The surrounding rock conditions were classified based on FPI and TPI, two key indices reflecting rock mass excavatability and rock-breaking efficiency. Using the K-means clustering algorithm, the dataset was segmented into three distinct groups. Seven tunneling parameters were selected as input features for the neural network model. Subsequently, three GA-BP neural network models were developed for different rock mass categories, with key parameters optimized for enhanced performance. Prediction results demonstrate that the GA-BP neural network exhibits superior accuracy and generalization capability. Compared to a conventional BP neural network, the GA-BP model reduces prediction errors by more than 10%.

1. Introduction

With the continuous advancement of tunneling technology, Tunnel Boring Machines (TBMs) have increasingly become the dominant method in underground construction, owing to their high degree of mechanization, enhanced safety, superior efficiency, minimal environmental impact, and reduced surface disturbances [1,2,3]. TBM advance speed is a critical metric for evaluating construction efficiency, directly influencing project timelines, costs, and quality [4]. However, the tunneling process involves complex geological conditions and multiple mechanical performance parameters, which exhibit highly nonlinear and high-dimensional interrelationships. These complexities make the accurate prediction of TBM advance speed a significant challenge in tunnel engineering.
Traditional methods for predicting TBM advance speed primarily rely on empirical formulas and statistical analyses. In 1977, Rosutami [5,6] pioneered the well-known CSM model through full-scale laboratory cutting tests, establishing relationships between rock properties, TBM thrust, penetration depth, and other factors. However, due to discrepancies between laboratory conditions and actual construction environments, the model’s conclusions tend to be conservative. The NTNU model [7], developed by the Norwegian University of Science and Technology, was constructed based on an extensive collection of geological parameters and tunneling data. While the incorporation of large-scale engineering data enhances its accuracy, the model suffers from limited generalization capability. Armaghani et al. [8] established a multiple regression equation linking geological parameters obtained from field observations and laboratory tests to TBM advance speed, proposing an empirical prediction method based on RQD, uniaxial compressive strength, and brittleness index. Jing et al. [9] incorporated variations in surrounding rock conditions into the influence of TBM parameters, constructing a predictive model using data collected from engineering sites. Nelson et al. [10] developed a precise TBM tunneling model by integrating detailed field-collected data. Through data processing and analysis, they built a probability-based model grounded in real-world data and further refined a penetration rate prediction model. While these methods capture certain trends in TBM advance speed, their inability to comprehensively account for the complex interactions between geological conditions and mechanical performance often limits their predictive accuracy and adaptability to engineering applications.
In recent years, with the rapid advancement of artificial intelligence, Artificial Neural Networks (ANNs) have demonstrated significant advantages in complex system modeling and prediction due to their powerful nonlinear mapping capabilities and self-learning characteristics [11,12,13]. This approach has also been widely applied in the prediction of TBM operational parameters [14,15,16,17]. Armaghani et al. [18] conducted a sampling survey on a tunnel project in Malaysia, establishing a machine learning dataset based on rock mechanics parameters and mechanical equipment data, and developed a TBM advance speed prediction model using ANN. Wang et al. [19] introduced an intelligent model based on a causal gated recurrent unit (GRU) to predict TBM performance parameters, which was validated with promising results. Gao et al. [20] incorporated the temporal characteristics of TBM data and constructed a recurrent neural network (RNN) for real-time prediction of TBM operational parameters. Meanwhile, a wide range of hybrid models integrated with optimization algorithms such as PSO, GA, EO, and DE have been extensively applied [21,22,23]. Fu et al. [24] developed a model combining a Genetic Algorithm with a BP neural network to predict the advance rate of Earth Pressure Balance shield tunneling. Grima et al. [25] developed a fuzzy neural network to predict TBM tunneling performance and compared it with other algorithms, demonstrating its strong predictive capability. Although machine learning has achieved some success in TBM advance speed prediction, several challenges remain. The excavation process of TBM is governed by the complex interaction between rock and machine. Most current models fail to integrate geological data, making it difficult to accurately predict TBM excavation speed under varying geological conditions. Additionally, most predictive methods primarily focus on exploring different algorithms, without deeply analyzing existing databases or considering the impact of data characteristics on prediction accuracy, leading to overfitting issues and insufficient prediction precision.
To address the challenges of predicting TBM advance speed under varying geological conditions and the complexity of redundant data. This study collected TBM construction data from a large-scale hydropower tunnel project in China. The dataset underwent extensive preprocessing and classification, resulting in three distinct subsets corresponding to different surrounding rock conditions. A correlation analysis was conducted on TBM operational parameters to eliminate redundant variables, thereby enhancing computational efficiency and model accuracy. For each dataset, a BP neural network model enhanced with a genetic algorithm (GA-BP model) was developed, with structural and parameter optimization performed to improve predictive performance. Consequently, TBM advance speed prediction models were established for varying geological conditions. Finally, the model’s effectiveness was validated using a test dataset.

2. Methods

2.1. K-Means Algorithm

K-means is a classical partition-based clustering algorithm, whose core concept involves the following: First, the number of clusters and the initial cluster centers are determined based on the characteristics of the study objects. Then, through an iterative process, the cluster centers are continuously adjusted to minimize the deviation between each data point and its corresponding cluster center [26]. Figure 1 illustrates an example of the K-means clustering algorithm. The specific steps of the K-means algorithm are as follows:
(1)
Randomly select k initial samples a 1 , a 2 , a 3 , …, a k as the cluster centers, corresponding to the clusters c 1 , c 2 , c 3 , …, c k .
(2)
For each sample x i in the dataset, calculate its distance to the k cluster centers and assign it to the group corresponding to the nearest center.
(3)
For each group c j , calculate its cluster center, which is the centroid of all the samples in that group, using the following formula:
a j = 1 c j x i c j x i
where c j represents the number of samples in group c j .
(4)
Calculate the error between the cluster centers before and after the update using the following formula:
E = j = 1 k a j ( t + 1 ) a j ( t ) 2
where E is the clustering criterion function, and a j ( t + 1 ) and a j ( t ) represent the cluster centers of group j in the (t + 1) and t iterations, respectively.
(5)
Repeat steps (2) to (4) until E converges to E m a x , where E m a x is the allowed maximum error.

2.2. BP Neural Network

The BP (Back-propagation) neural network is a multi-layer feedforward network based on error backpropagation, known for its simple structure, strong nonlinear mapping ability, high fault tolerance, and excellent generalization capability. It has been widely applied across various fields such as industry, agriculture, and medicine, making it one of the most mature and frequently used algorithms. Figure 2 illustrates the structure of a BP neural network, which includes an input layer, hidden layers, and an output layer. The input layer is responsible for receiving external input data, with each input node representing one dimension of the input features. The hidden layer typically consists of one or more layers, each containing multiple neurons. The function of the neurons in the hidden layer is to perform nonlinear transformations on the input data, allowing the network to learn complex mapping relationships. The output layer generates the prediction results, with the number of neurons varying depending on the specific task. Additionally, each neuron has an activation function, commonly Sigmoid, Tanh, or Relu [27]. The activation function introduces nonlinearity, enabling the neural network to approximate complex functions.
During the training process, the BP neural network adjusts its parameters through two main steps: forward propagation and backpropagation. In the forward propagation phase, input data starts from the input layer and passes through each layer, where a weighted sum and activation function are applied, as shown in Equation (3), ultimately reaching the output layer to generate the predicted result. Specifically, each neuron receives input from the neurons of the previous layer, computes the weighted sum, adds a bias, and then undergoes a nonlinear transformation via the activation function before passing the result to the next layer. This way, the network performs layer-by-layer calculations, ultimately producing a predicted value.
μ i , k = f ( i = 1 n w i j , k · x i b k )
where f( ) represents the activation function, which introduces nonlinearity into the neural network’s fitting process. μi,k is the output value of the i-th neuron in the k-th layer; bk is the bias of the k-th layer; xi is the i-th input; and wij,k represents the connection weight between the i-th input of the k-th layer and the j-th neuron of the next layer.
The second part is the backpropagation of the error information. When the difference between the computed result and the true value does not meet the required criteria, the error is propagated backward along the same path as the forward pass. During this process, the connection weights and bias values of each neuron are adjusted to reduce the error. This cycle continues iteratively until the error meets the desired threshold or the maximum number of iterations is reached.

2.3. Genetic Algorithm

The core idea of the genetic algorithm [28] is to optimize target parameters by simulating the biological evolution process. It encodes numerous feasible solutions of the target parameters to form an initial chromosome, which serves as the basic unit of the population’s evolution. Using a fitness function as a measure, genetic operations such as selection, crossover, and mutation are applied to recombine and optimize the coding genes of different chromosomes within the population. This results in the generation of a new offspring population, establishing an iterative process where the fitness of the offspring is aimed to surpass that of the parents, thus achieving the overall evolution of the population and progressively approaching the optimal solution for the target parameters. The basic process of the genetic algorithm is as follows:
(1)
Initialize the population: The genetic algorithm encodes the feasible solutions of the target parameters into chromosomes based on the established encoding mechanism, and then integrates individual chromosomes to form the initial population.
(2)
Evaluate the fitness of the parent population: The optimization objective of the parameters is used as the fitness function, and the fitness values of the individuals in the population are calculated to assess their quality.
(3)
Genetic operations: Genetic operations include selection, crossover, and mutation. Selection refers to the process of probabilistically choosing individuals with high fitness from the parent population to be part of the offspring population. Crossover refers to the process of probabilistically exchanging certain gene segments between two selected chromosomes to create new individuals that become part of the offspring population. Mutation refers to the process of probabilistically changing certain genes in a chromosome from 0 to 1 (or vice versa). The purpose of crossover and mutation is to increase population diversity.
(4)
Evaluate the fitness of the offspring population: Observe and calculate whether the fitness values of the offspring population meet the accuracy requirements. If the requirements are met, the algorithm terminates. Otherwise, genetic operations are repeated, and the iteration continues until the requirements are met or the maximum number of iterations is reached.
Although BP neural networks have significant advantages and great development potential, they also face issues in practical production applications, such as slow convergence, low efficiency, and a tendency to get trapped in local optima. On the other hand, the genetic algorithm, as a global optimization search algorithm, can locate the global optimum based solely on the evaluation function (fitness function). Therefore, the genetic algorithm can be used to optimize the BP neural network. In this study, the genetic algorithm is applied to optimize the hyperparameters of the BP network (i.e., the initial weights and biases), constructing a neural network with strong nonlinear mapping ability, good learning performance, and global search capabilities. This approach addresses the shortcomings of the BP neural network, enhances its performance, and achieves a synergistic effect through the combination of both techniques. Figure 3 illustrates the computation process of the entire GA-BP model.

3. Data Processing and Analysis

This study is based on the exploratory tunnel project of a large hydropower station in China, encompassing TBM data, parameters, and other essential information. The data spans approximately fifteen months of construction (from July 2023 to October 2024), primarily consisting of parameters generated by sensors installed at key positions on the TBM.

3.1. Abnormal Data Process

On a typical workday, a single TBM records thousands of data points, each containing multiple tunneling parameters. This results in raw construction data that is voluminous and sparse in terms of information density. Therefore, it is essential to preprocess the engineering data to extract the necessary information.
During TBM operation, there are periods of non-working states, such as step changes, shutdowns, and maintenance, as illustrated in Figure 4. The tunneling data during these phases mostly consists of meaningless zero values. Consequently, it is necessary to first extract the working segment data. This can be carried out by identifying zero values in the tunneling parameters that characterize the working state, such as thrust, cutterhead rotation speed, cutterhead torque, and advance speed. The discrimination function for extracting the working segment is as follows:
g x = 1 , x > 0 0 , x 0
G = g F · g q · g T · g v
where G is the TBM tunneling state discrimination value, where G = 1 indicates tunneling state and G = 0 indicates non-tunneling state. F, q, T, and v represent thrust, cutterhead rotation speed, cutterhead torque, and advance speed, respectively.
Meanwhile, due to the maximum stroke of the TBM hydraulic cylinder, the ideal advance rate for a normal excavation is one that approaches this value as closely as possible. In the data, excavation cycles with relatively short advance distances (or durations that are too brief) typically occur due to special circumstances that cause stoppages, and these can be considered abnormal excavation cycles. In this project, a single TBM advance cycle requires a minimum of 20 min. Therefore, to ensure data quality, excavation cycles with a total duration of less than 20 min should be excluded. Additionally, tunnel construction is subject to various sources of noise interference, and due to sensor errors and human mistakes, the TBM operational data inevitably contains some outliers, as illustrated in Figure 5. This paper employs a boxplot method based on quartiles to eliminate outliers in the stable section of the data. In this method, the upper quartile Q 3 , lower quartile Q 1 , and interquartile range ( I Q R ) are used to determine the upper and lower boundaries ( L u p p e r and L l o w e r ) for non-outlier values. The specific criteria for outlier identification and removal are as follows.
L u p p e r = Q 3 + 1.5 I Q R
L l o w e r = Q 1 1.5 I Q R
I Q R = Q 3 Q 1
D s Abnormal   value   ( D s < L l o w e r   or   D s > L u p p e r ) Normal   value   ( L l o w e r < D s < L u p p e r )
After completing the preprocessing of the engineering data, a typical parameter variation curve for a full excavation day is shown in Figure 6. As seen in the figure, after construction begins, the typical TBM excavation parameters rise sharply and quickly stabilize. The trends of all parameters are similar, remaining consistent until the completion of the work for the day.

3.2. Distribution of TBM Parameters Based on FPI and TPI

The excavation rate of the TBM varies significantly under different operating conditions, with geological conditions and TBM performance being the primary influencing factors. However, in practical engineering, it is difficult to obtain real-time geological conditions of the excavation face. Therefore, this paper refers to Wang et al. [29], which classifies the dataset using comprehensive characteristic indicators of excavation rock breaking, including Torque per Incision (TPI) and Force per Incision (FPI). These composite indicators help mitigate the influence of human operational factors to some extent and primarily reflect the rock excavatability of the surrounding rock and the rock breaking efficiency of the TBM at the current excavation cycle location. The definitions of TPI and FPI are as follows:
T P I = F P r
F P I = T P r
The FPI is defined as the total thrust force (F) divided by the penetration rate (Pr), while the TPI is the cutterhead torque (T) divided by the penetration rate. A higher TPI value indicates greater shear strength of the surrounding rock and increased cutterhead rotational resistance, whereas a lower TPI value suggests softer surrounding rock with reduced rotational resistance. A higher FPI value signifies harder surrounding rock and greater excavation difficulty, while a lower FPI value indicates softer surrounding rock and higher excavation efficiency.
Using the above formulas, the TPI and FPI of the dataset can be calculated. Subsequently, the K-means clustering algorithm described in Section 2.1 is applied to categorize the data under different operating conditions. As shown in Figure 7a, the dataset is divided into three categories, each corresponding to a different type of surrounding rock. Dataset 1 contains 9163 samples, Dataset 2 includes 24,495 samples, and Dataset 3 comprises 44,903 samples. The means of the three datasets are 2.41, 7.63, and 35.52, with standard deviations of 0.97, 2.87, and 15.26, respectively. Figure 7b–d illustrate the distribution characteristics of advance speed across the three datasets, clearly revealing that the advance speed does not follow a normal distribution. As the surrounding rock strength decreases, a significant increase in advance speed is observed.

3.3. Analysis of the Impact of TBM Parameters on Advance Speed

The various sensors installed on the TBM are capable of recording hundreds of operational parameters throughout the excavation process. However, not all data (such as hydraulic oil tank temperature, oxygen content, etc.) are useful for constructing subsequent excavation rate prediction models. In extreme cases, such data may even reduce model training efficiency and predictive accuracy. Therefore, conducting a correlation analysis to identify the most influential excavation parameters on the excavation rate is crucial, as it provides a solid foundation for future work and holds significant importance.
In this context, this paper selects 13 parameters from the over 100 excavation parameters recorded by the sensors installed on the TBM, focusing on those related to the rock breaking system, propulsion, and support system. Using construction data from a full excavation cycle, the Pearson correlation coefficient is employed to calculate the correlation between the excavation speed and other parameters. These parameters are then further filtered to identify the optimal input variables, aiming to reduce data dimensionality, simplify the model, and enhance prediction efficiency and accuracy. The Pearson correlation coefficient can be calculated using the following formula:
ρ X , Y = c o v ( X , Y ) σ ( X ) σ ( Y )
where c o v ( X , Y ) is the covariance between the random variables X and Y, and σ ( X ) and σ ( Y ) represent the standard deviations of the variables. If X and Y are independent, then ρ X , Y = 0 . The correlation coefficient ranges from 0 to 1, indicating a transition from no correlation to strong correlation between the two variables.
The correlation results are shown in Figure 8. According to Hair et al. [30], an absolute Pearson correlation coefficient greater than 0.4 is considered to indicate a moderate correlation. To ensure comprehensive information and efficient training, tunneling parameters with Pearson correlation coefficients exceeding 0.4 in absolute value with respect to TBM advance rate were selected. These include TBM inclination (−0.64944), penetration rate (0.8414), total thrust (0.67703), cutterhead speed (0.78825), cutterhead torque (0.5284), propulsion pump pressure (0.67902), and main belt conveyor drive pressure (−0.45282). These seven parameters were chosen as input features for the ANN model.

4. Application of GA-BP Neural Network Model

For the three datasets corresponding to the three types of surrounding rock, this paper constructs three artificial neural network models to adapt to different rock conditions. The following sections will provide a detailed description of the model construction, optimization, and performance evaluation. In addition, all processes including data processing and the construction, training, and evaluation of neural network models were completed on a computer equipped with an i5-10200 CPU and a GTX1650Ti GPU.

4.1. Model Building and Optimization

Three BP neural network models were constructed as shown in Figure 2. Each model includes an input layer (with 7 neurons), one hidden layer, and an output layer (with 1 neuron). The number of neurons in the input layer is determined by the number of excavation parameters selected in Section 3.3. The number of neurons in the hidden layer is subject to optimization, with the output representing the advance speed. The activation function in the hidden layer is the ReLU activation function, while the activation function in the output layer is the tanh activation function.
t a n h ( x ) = e x e x e x + e x
r e l u ( x ) = x ,     for   x 0 0 ,     for   x < 0
In the BP neural network, the number of neurons in the hidden layer has a significant impact on the overall performance of the model. Therefore, this section uses an experimental approach to determine the optimal number of hidden layer neurons. A total of 5000 data samples were selected from each of the three datasets. Given the differing dimensions and value ranges of the various parameters, a considerable burden is placed on improving the accuracy and efficiency of model prediction. To address this, min-max normalization [22], as defined in Equation (15), was applied to the parameters before they were used as model inputs.
x = x x m i n x m a x x m i n
In determining the number of hidden layer neurons, the mean absolute error (MAE) is set as the loss function, the batch size is set to 32, and each model is trained for 1000 iterations. Due to the differences in the value ranges of the three datasets, the learning rates for the three models are set to 0.01, 0.005, and 0.001, respectively, with the Adam optimizer [31] used. The range for the number of hidden layer neurons is set from 8 to 32. Figure 9 shows the impact of the number of hidden layer neurons on the prediction error of the models. From the figure, it is evident that the number of hidden layer neurons has a significant effect on the model’s prediction accuracy. As the number of neurons increases, the MAE initially decreases and then increases. For dataset 1, the MAE reaches its minimum value of 0.36 when the number of neurons is 12. For dataset 2, the MAE also reaches its minimum value of 0.48 when the number of neurons is 12. For dataset 3, the MAE reaches its minimum value of 2.42 when the number of neurons is 24.
When introducing a genetic algorithm into the BP neural network for optimizing the initial weights and thresholds, the population size directly affects the algorithm’s final prediction performance. If the population size is too small, it cannot fully represent the population information, potentially leading to premature convergence and local extremum issues. On the other hand, if the population size is too large, while the algorithm’s accuracy improves, its computational efficiency will decrease. Therefore, in this section, the population sizes are set to 20, 30, 40, 50, 60, 70, 80, 90, and 100 to identify the optimal parameters. Additionally, the crossover probability is set to 0.7, and the mutation probability is set to 0.03. The training results are shown in Figure 10, where it can be observed that as the population size increases, the MAE for the three models fluctuates in a wave-like pattern. For the three datasets, the MAE reaches its minimum value when the population sizes are 60, 60, and 50, respectively. In conclusion, the optimal structural parameters for the GA-BP neural network excavation speed prediction model established in this paper are shown in Table 1.

4.2. Evaluation of the Model Performance

Next, the optimized models are used to train the complete datasets for the three types of surrounding rock. Following Zhong et al. [32], each of the three datasets was divided into 80% for training and 20% for testing.
To provide a quantitative analysis of the model’s prediction performance, this paper uses the Mean Absolute Error (MAE), Root Mean Squared Error (RMSE), and R-squared (R2) as evaluation metrics to assess the overall performance of the model. The formulas for calculating these metrics are as follows:
R 2 = i = 1 n ( y i y ¯ ) 2 i = 1 n ( y i y ¯ ) 2
R M S E = i = 1 n ( y i y i ) 2 n
where y i and y i represent the actual and predicted values for the i-th sample, y ¯ is the mean of the actual values, and n is the number of test samples.
Figure 11 shows the prediction results of the GA-BP neural network on the training set. It can be observed that the GA-BP neural network performs well on all three datasets, with R2 values above 0.9. Additionally, as the surrounding rock conditions transition from hard to soft, the prediction results of the GA-BP neural network become increasingly accurate. This may be related to the size of the datasets. In this project, the TBM encounters fewer tunnel sections with hard surrounding rock during excavation, and most tunnel sections have more fractured surrounding rock. Overall, the three GA-BP neural networks developed in this study are capable of predicting the TBM advance speed under different surrounding rock conditions.
Since the test set was not involved in the training process, the trained GA-BP neural network was used to predict the excavation speed on the test set to evaluate the model’s generalization ability. Figure 12 shows the prediction results for the test set. It can be observed that under the conditions of the three surrounding rock types, the R2 values reached 0.91, 0.94, and 0.93, respectively. These results demonstrate that the model possesses strong predictive performance and a certain level of generalization capability.
Meanwhile, the original BP neural network and convolutional neural network (CNN) model were employed to predict the test set. Figure 13 compares the predictive performance of the conventional BP model, CNN model and the GA-BP model across three test sets. As shown in Figure 13b,e,h, the BP model can partially capture the overall trend of advance speed but performs poorly in tracking sudden shifts and local peaks, with pronounced deviations particularly in high-frequency fluctuation regions. For instance, during phases of rapid acceleration or deceleration, the BP model generally underestimates or lags behind actual values. This indicates that the BP model is prone to local optima during training, with limited capacity to fit highly nonlinear features. Additionally, the CNN model (Figure 13c,f,i) demonstrates relatively stable performance in predicting overall trends and effectively reflects the changes in advance speed. However, it still shows some deviation when handling local high-frequency fluctuations. Notably, in Dataset 3, its prediction performance is less stable than that of the GA-BP model. In contrast, the GA-BP model (Figure 13a,d,g) demonstrates superior predictive accuracy. By globally optimizing the initial weights and thresholds of the BP model, the genetic algorithm enhances its global search capability, enabling more precise fitting of the nonlinear dynamic characteristics of advance speed. The figures clearly show that GA-BP predictions align more closely with measured data, maintaining robust tracking even under sharp fluctuations and abrupt changes. The precise prediction capability of the GA-BP model enables the project team to more effectively schedule construction progress, adjust work plans, and optimize resource allocation. This maximizes construction efficiency while avoiding excessive labor input or equipment idling. As construction progresses, the model dynamically updates its predictions, providing real-time feedback to the project team. This helps them better adapt to potential changes during the construction process, ensuring that both construction efficiency and cost control remain optimized. Overall, the application of the GA-BP neural network model in TBM advance speed prediction not only enhances the management level of engineering projects but also effectively controls construction costs, reduces project risks, and improves the economic benefits and feasibility of project implementation.
Table 2 further quantifies the predictive performance of the different models. Comparative analysis reveals that, relative to the original BP neural network, the model optimized through the genetic algorithm exhibits consistent improvements across all three datasets. Specifically, the MAE decreases by 21.95%, 13.46%, and 14.19%, while the RMSE is reduced by 16.27%, 10.53%, and 11.18%. And the coefficient of determination shows a notable enhancement. The CNN model also outperforms the original BP neural network in prediction accuracy. Furthermore, compared to the CNN model, the GA-BP model shows reductions in MAE by 15.79%, 8.16%, and 11.43%, and in RMSE by 12.20%, 8.93%, and 8.99% across the three datasets. These results provide compelling evidence of the effectiveness and accuracy of the proposed predictive model.

5. Conclusions

Based on a large amount of actual TBM construction data, this study removed non-operational state data and anomalous values to create a reasonable TBM construction parameter database. It then used the FPI and TPI composite feature indicators, along with the K-means fuzzy clustering method, to classify the surrounding rock, dividing the dataset into three parts. Subsequently, for the three surrounding rock conditions, three GA-BP neural network models were established to predict the TBM advance speed. The main conclusions are as follows:
(1)
Through correlation analysis, seven TBM construction parameters that significantly impact advance speed were selected, including TBM inclination, penetration rate, total thrust, cutterhead speed, cutterhead torque, propulsion pump pressure, and main belt conveyor drive pressure.
(2)
For the three different surrounding rock conditions, the three BP neural network models optimized by experimental methods and genetic algorithms all demonstrated high prediction accuracy, effectively reflecting the relationship between key construction parameters and excavation speed.
(3)
By calculating model performance indicators, the GA-BP neural network demonstrates superior predictive accuracy and generalization capability compared to the original BP neural network. Under three surrounding rock conditions ranging from hard to soft, the GA-BP model achieves reductions in MAE by 21.95%, 13.46%, and 14.19%, and in RMSE by 16.27%, 10.53%, and 11.18%, respectively.
This study has completed an initial prediction of TBM advance speed. However, there are still some limitations. Due to constraints in the field conditions, the mechanical properties of the surrounding rock were not obtained in real time. In future research, rock samples from different tunnel sections can be collected, and the characteristic parameters of the surrounding rock can be incorporated into the neural network model to further enhance the model’s accuracy and generalization capabilities.

Author Contributions

W.-F.Z.: Resources, Investigation. S.-Q.C.: Resources, Investigation. J.-W.Z.: Writing—review and editing, Investigation, Conceptualization. X.-F.W.: Resources, Investigation. Y.-H.R.: Resources, Writing—review and editing. H.-B.L.: Writing—review and editing. Z.-Q.W.: Resources, Investigation. B.N.: Writing—original draft, Software, Methodology, Formal analysis. All authors have read and agreed to the published version of the manuscript.

Funding

This study was supported by the Sichuan Youth Science and Technology Innovation Research Team Project (2020JDTD0006) and the Open Research Fund of Key Laboratory of Reservoir and Dam Safety Ministry of Water Resources (YK323002).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets used during the current study available from the corresponding author on reasonable request.

Conflicts of Interest

Authors Wei-Feng Zhang, Shi-Quan Chen, Xiang-Feng Wang and Zhi-Qiang Wang were employed by the company Power China Chengdu Engineering Corporation Limited. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Zare, S.; Bruland, A. Applications of MTNU/SINTEF drillability indices in hard rock tunneling. Rock Mech. Rock Eng. 2013, 46, 179–187. [Google Scholar] [CrossRef]
  2. Delisio, A.; Zhao, J. A new model for TBM performance prediction in blocky rock conditions. Tunn. Undergr. Space Technol. 2014, 43, 440–452. [Google Scholar] [CrossRef]
  3. Huo, J.; Sun, W.; Chen, J.; Zhang, X. Disc cutters plane layout design of the full-face rock tunnel boring machine (TBM) based on different layout patterns. Comput. Ind. Eng. 2011, 61, 1209–1225. [Google Scholar] [CrossRef]
  4. Hassanpour, J.; Rostami, J.; Zhao, J. A new hard rock TBM performance prediction model for project planning. Tunn. Undergr. Space Technol. 2011, 26, 595–603. [Google Scholar] [CrossRef]
  5. Rosutami, J. A new model for performance prediction of hard rock TBMs. In Proceedings of the 1993 Rapid Excavation and Tunneling Conference, Boston, MA, USA, 13–17 June 1993. [Google Scholar]
  6. Rosutami, J. Development of a Force Estimation Model for Rock Fragmentation with Disc Cutters Through Theoretical Modeling and Physical Measurement of Crushed Zone Pressure; Colorado School of Mines: Golden, CO, USA, 1997. [Google Scholar]
  7. Bruland, A. Hard Rock Tunnel Boring; Fakultet for Ingeniørvitenskap og Teknologi: Trondheim, Norway, 2000. [Google Scholar]
  8. Armaghani, D.; Faradonbeh, R.; Momeni, E.; Fahimifar, A.; Tahir, M.M. Performance prediction of tunnel boring machine through developing a gene expression programming equation. Eng. Comput. 2018, 34, 129–141. [Google Scholar] [CrossRef]
  9. Jing, L.; Li, J.; Zhang, N.; Chen, S.; Yang, C.; Cao, H. A TBM advance rate prediction method considering the effects of operating factors. Tunn. Undergr. Space Technol. 2021, 107, 103620. [Google Scholar] [CrossRef]
  10. Nelson, P.; Rourke, D.; Kulhawy, H. Factors affecting TBM penetration rates in sedimentary rocks. In Proceedings of the 24th US Symposium on Rock Mechanics, College Station, TX, USA, 20–23 June 1983; pp. 227–237. [Google Scholar]
  11. Bardhana, K.; Kardanin, G.; Guharay, A.; Burman, A.; Samui, P.; Zhang, Y. Hybrid ensemble soft computing approach for predicting penetration rate of tunnel boring machine in a rock environment. J. Rock Mech. Geotech. Eng. 2021, 13, 1398–1412. [Google Scholar] [CrossRef]
  12. Zhou, J.; Qiu, Y.G.; Zhu, S.L.; Armaghani, D.J.; Li, C.; Nguyen, H.; Yagiz, S. Optimization of support vector machine through the use of metaheuristic algorithms in forecasting TBM advance rate. Eng. Appl. Artif. Intell. 2021, 97, 104015. [Google Scholar] [CrossRef]
  13. Wang, X.; Zhu, H.; Zhu, M.; Zhang, L.; Ju, J.W. An integrated parameter prediction framework for intelligent TBM excavation in hard rock. Tunn. Undergr. Space Technol. 2021, 118, 104196. [Google Scholar] [CrossRef]
  14. Benardos, A.G.; Kaliampakos, D.C. Modelling TBM performance with artificial neural networks. Tunn. Undergr. Space Technol. 2004, 19, 597–605. [Google Scholar] [CrossRef]
  15. Yagiz, S.; Karahan, H. Prediction of hard rock TBM penetration rate using particle swarm optimization. Int. J. Rock Mech. Min. Sci. 2011, 48, 427–433. [Google Scholar] [CrossRef]
  16. Salimi, A.; Rostami, J.; Moormann, C.; Delisio, A. Application of non-linear regression analysis and artificial intelligence algorithms for performance prediction of hard rock TBMs. Tunn. Undergr. Space Technol. 2016, 58, 236–246. [Google Scholar] [CrossRef]
  17. Adoko, A.C.; Gokceoglu, C.; Yagiz, S. Bayesian prediction of TBM penetration rate in rock mass. Eng. Geol. 2017, 226, 245–256. [Google Scholar] [CrossRef]
  18. Armaghani, D.J.; Mohamad, E.T.; Narayanasamy, M.S.; Narita, N.; Yagiz, S. Development of hybrid intelligent models for predicting TBM penetration rate in hard rock conditions. Tunn. Undergr. Space Technol. 2017, 63, 29–43. [Google Scholar] [CrossRef]
  19. Wang, K.; Zhang, L.; Fu, X. Time series prediction of tunnel boring machine (TBM) performance during excavation using causal explainable artificial intelligence (CX-AI). Autom. Constr. 2023, 147, 104730. [Google Scholar] [CrossRef]
  20. Gao, X.; Shi, M.; Song, X.; Zhang, C.; Zhang, H. Recurrent neural networks for real-time prediction of TBM operating parameters. Autom. Constr. 2019, 98, 225–235. [Google Scholar] [CrossRef]
  21. Pradeep, T.; Samui, P. Prediction of rock strain using hybrid approach of ANN and optimization algorithms. Geotech. Geol. Eng. 2022, 40, 4617–4643. [Google Scholar] [CrossRef]
  22. Rahul, B.; Manish, K.; Divesh, R.K.; Samui, P.; Pradeep, T.; Rajak, M.K.; Armaghani, D.J.; Singh, S. Application of novel deep neural network on prediction of compressive strength of fly ash based concrete. Nondestruct. Test. Eval. 2024, 40, 4638–4668. [Google Scholar]
  23. Abdul, B.J.; Pradeep, T. ANN-PSO modelling for predicting buckling of self-compacting concrete column containing RHA properties. Matéria 2023, 28, e20230102. [Google Scholar]
  24. Fu, X.S.; Gong, Q.M.; Wu, Y.J.; Zhao, Y.; Li, H. Prediction of EPB shield tunneling advance rate in mixed ground condition using optimized BPNN model. Appl. Sci. 2022, 12, 5485. [Google Scholar] [CrossRef]
  25. Grima, M.A.; Bruines, P.A.; Verhoef, P.N.W. Modeling tunnel boring machine performance by neuro-fuzzy methods. Tunn. Undergr. Space Technol. 2000, 15, 259–269. [Google Scholar] [CrossRef]
  26. Dubey, A.; Choubey, A. A systematic review on k-means clustering techniques. Int. J. Sci. Res. Eng. Technol. 2017, 6, 1456–1475. [Google Scholar]
  27. Yarotsky, D. Error bounds for approximations with deep ReLU networks. Neural Netw. 2017, 94, 103–114. [Google Scholar] [CrossRef]
  28. Mirjalili, S. Genetic Algorithm. In Evolutionary Algorithms and Neural Networks; Springer: Cham, Switzerland, 2019; p. 780. [Google Scholar]
  29. Wang, H.; Wang, J.; Zhao, Y.; Xu, H. Tunneling parameters optimization based on multi-objective differential evolution algorithm. Soft Comput. 2021, 25, 3637–3656. [Google Scholar] [CrossRef]
  30. Hair, J.F.; Anderson, R.E.; Tatham, R.L.; Black, W.C. Multivariate Data Analysis; Prentice Hall Iberia: Madrid, Spain, 2007. [Google Scholar]
  31. Kingma, D.P.; Ba, J.L. Adam: A method for stochastic optimization. Comput. Res. Repos. 2014, 1412, 74–89. [Google Scholar]
  32. Zhong, Z.L.; Ni, B.; Shen, J.X.; Du, X.L. Neural network prediction model for site response analysis based on the KiK-net database. Comput. Geotech. 2024, 171, 106366. [Google Scholar] [CrossRef]
Figure 1. Example of K-means algorithm (k = 3).
Figure 1. Example of K-means algorithm (k = 3).
Applsci 15 12115 g001
Figure 2. BP neural network structure.
Figure 2. BP neural network structure.
Applsci 15 12115 g002
Figure 3. GA-BP neural network computation process.
Figure 3. GA-BP neural network computation process.
Applsci 15 12115 g003
Figure 4. TBM working and non-working states.
Figure 4. TBM working and non-working states.
Applsci 15 12115 g004
Figure 5. Abnormal values during the excavation process.
Figure 5. Abnormal values during the excavation process.
Applsci 15 12115 g005
Figure 6. Variation curve of typical TBM parameters during an excavation day.
Figure 6. Variation curve of typical TBM parameters during an excavation day.
Applsci 15 12115 g006
Figure 7. Distribution characteristics of the datasets. (a) Clustering results based on TPI and FPI; (b) Distribution of advance speed in dataset 1; (c) Distribution of advance speed in dataset 2; (d) Distribution of advance speed in dataset 3.
Figure 7. Distribution characteristics of the datasets. (a) Clustering results based on TPI and FPI; (b) Distribution of advance speed in dataset 1; (c) Distribution of advance speed in dataset 2; (d) Distribution of advance speed in dataset 3.
Applsci 15 12115 g007
Figure 8. Correlation coefficients between various parameters and advance speed.
Figure 8. Correlation coefficients between various parameters and advance speed.
Applsci 15 12115 g008
Figure 9. MAE for different number of nodes in the hidden layer.
Figure 9. MAE for different number of nodes in the hidden layer.
Applsci 15 12115 g009
Figure 10. MAE under different population sizes.
Figure 10. MAE under different population sizes.
Applsci 15 12115 g010
Figure 11. Prediction results for the training set. (a) Dataset 1; (b) Dataset 2; (c) Dataset 3.
Figure 11. Prediction results for the training set. (a) Dataset 1; (b) Dataset 2; (c) Dataset 3.
Applsci 15 12115 g011
Figure 12. Prediction results for the test set. (a) Dataset 1; (b) Dataset 2; (c) Dataset 3.
Figure 12. Prediction results for the test set. (a) Dataset 1; (b) Dataset 2; (c) Dataset 3.
Applsci 15 12115 g012
Figure 13. Comparison of advance speed predicted by different models. (a) Prediction by GA-BP model in dataset 1; (b) Prediction by BP model in dataset 1; (c) Prediction by CNN model in dataset 1; (d) Prediction by GA-BP model in dataset 2; (e) Prediction by BP model in dataset 2; (f) Prediction by CNN model in dataset 2; (g) Prediction by GA-BP model in dataset 3; (h) Prediction by BP model in dataset 3; (i) Prediction by CNN model in dataset 3.
Figure 13. Comparison of advance speed predicted by different models. (a) Prediction by GA-BP model in dataset 1; (b) Prediction by BP model in dataset 1; (c) Prediction by CNN model in dataset 1; (d) Prediction by GA-BP model in dataset 2; (e) Prediction by BP model in dataset 2; (f) Prediction by CNN model in dataset 2; (g) Prediction by GA-BP model in dataset 3; (h) Prediction by BP model in dataset 3; (i) Prediction by CNN model in dataset 3.
Applsci 15 12115 g013
Table 1. Optimal model parameters.
Table 1. Optimal model parameters.
ParametersModel 1Model 2Model 3
Number of neurons in hidden layer121224
Learning rate0.0010.0050.01
Batch size323232
Population size406060
Crossover probability0.70.70.7
Mutation probability0.030.030.03
Table 2. Evaluation indicators for the prediction results of different models.
Table 2. Evaluation indicators for the prediction results of different models.
ModelDataset 1Dataset 2Dataset 3
MAERMSER2MAERMSER2MAERMSER2
GA-BP0.320.360.910.450.510.942.544.450.93
BP0.410.430.880.520.570.902.965.010.91
CNN0.380.410.890.490.560.912.874.890.91
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, W.-F.; Chen, S.-Q.; Zhou, J.-W.; Wang, X.-F.; Ran, Y.-H.; Li, H.-B.; Wang, Z.-Q.; Ni, B. Hybrid GA-BP Neural Network for Accurate Prediction of TBM Advance Speed Under Complex Geological Conditions. Appl. Sci. 2025, 15, 12115. https://doi.org/10.3390/app152212115

AMA Style

Zhang W-F, Chen S-Q, Zhou J-W, Wang X-F, Ran Y-H, Li H-B, Wang Z-Q, Ni B. Hybrid GA-BP Neural Network for Accurate Prediction of TBM Advance Speed Under Complex Geological Conditions. Applied Sciences. 2025; 15(22):12115. https://doi.org/10.3390/app152212115

Chicago/Turabian Style

Zhang, Wei-Feng, Shi-Quan Chen, Jia-Wen Zhou, Xiang-Feng Wang, Yu-Han Ran, Hai-Bo Li, Zhi-Qiang Wang, and Bo Ni. 2025. "Hybrid GA-BP Neural Network for Accurate Prediction of TBM Advance Speed Under Complex Geological Conditions" Applied Sciences 15, no. 22: 12115. https://doi.org/10.3390/app152212115

APA Style

Zhang, W.-F., Chen, S.-Q., Zhou, J.-W., Wang, X.-F., Ran, Y.-H., Li, H.-B., Wang, Z.-Q., & Ni, B. (2025). Hybrid GA-BP Neural Network for Accurate Prediction of TBM Advance Speed Under Complex Geological Conditions. Applied Sciences, 15(22), 12115. https://doi.org/10.3390/app152212115

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop