Next Article in Journal
A Lightweight Traffic Sign Small Target Detection Network Suitable for Complex Environments
Previous Article in Journal
BIM-Based Machine Learning Framework for Early-Stage Building Energy Performance Prediction
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Shield Thrust Time-Series Prediction Based on BiLSTM with Intelligent Hyperparameter Optimization

1
School of Water Resources and Hydroelectric Engineering, North China Electric Power University, Beijing 102206, China
2
Beijing Construction Engineering Group Co., Ltd., Beijing 100055, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2026, 16(1), 325; https://doi.org/10.3390/app16010325
Submission received: 15 December 2025 / Revised: 25 December 2025 / Accepted: 27 December 2025 / Published: 28 December 2025

Abstract

Shield thrust is a key control parameter for ensuring the safety and efficiency of tunnel construction. Under complex geological conditions and strong data nonlinearity, conventional prediction methods often fail to achieve sufficient accuracy. This study proposes a hybrid prediction model in which a bidirectional long short-term memory (BiLSTM) network is optimized by intelligent algorithms. A multidimensional input dataset comprising tunnel geometry, geomechanical parameters and tunnelling parameters is constructed, and BiLSTM is used to capture bidirectional temporal dependencies in the tunnelling data. To adaptively determine key BiLSTM hyperparameters (number of neurons, dropout rate and learning rate), four intelligent optimization algorithms—genetic algorithm (GA), particle swarm optimization (PSO), sparrow search algorithm (SSA) and Hunger Games Search (HGS)—are employed for hyperparameter tuning, using the root-mean-square error (RMSE) between predicted and measured values as the fitness function. Four hybrid models, GA–BiLSTM, PSO–BiLSTM, SSA–BiLSTM and HGS–BiLSTM, are validated using real engineering data from Beijing Metro Line 22, Guanzhuang–Yongshun shield-driven section. The results show that HGS–BiLSTM outperforms the other models in terms of RMSE, mean absolute error (MAE), mean absolute percentage error (MAPE) and coefficient of determination (R2) and exhibits faster convergence, supporting real-time prediction and decision-making in shield thrust control.

1. Introduction

With the in-depth implementation of new-type urbanization and the strategy of building a strong transportation power, China’s urban spatial pattern is undergoing profound restructuring. Faced with scarce land resources and saturated existing transportation systems, developing a high-capacity, high-density underground urban rail transit network has become an inevitable choice to break through urban development bottlenecks and optimize the spatial structure. In this systemic undertaking, advances in tunnel construction technology directly determine construction efficiency, safety and environmental performance.
Among various methods such as cut-and-cover and shallow tunnelling, the shield tunnelling method, with its “minimally invasive” ability to pass beneath cities, offers high safety, low environmental impact and good economy. It is evolving from an advanced construction technique into the “backbone” and core driver of efficient underground space development in modern cities [1,2,3].
The thrust of a shield machine is the resultant force applied by the machine to overcome the resistance of the soil at the tunnel face and the frictional resistance of the lining segments, thereby driving the cutterhead and segments forward during tunnelling. It is a key control parameter that reflects both the tunnelling capacity of the shield machine and the stability of the tunnel face. Insufficient thrust may lead to risks such as face instability, over-excavation and even collapse, whereas excessive thrust tends to increase ground disturbance and additional stress, potentially inducing surface settlement, deformation of adjacent buildings and damage to lining segments [4]. Therefore, under complex geological conditions and diverse operating conditions, accurately predicting and optimally controlling shield thrust is of great significance for ensuring tunnelling safety and maintaining the stability of the surrounding environment. Hereafter, the total shield thrust is denoted by TH.
During shield tunnelling, sudden changes in ground conditions and complex interactions among construction parameters make it difficult to predict variations in shield thrust in a timely and reliable manner based solely on operator experience. Earlier studies mainly used empirical formulas or laboratory tests to investigate shield thrust. Krause et al. [5] proposed an empirical formula for estimating shield thrust based on multiple engineering cases in Japan and Germany. Owing to its simple form, it enables site engineers to quickly approximate and compare thrust demands. However, shield thrust is influenced by numerous factors, and this empirical formula considers shield diameter as the main independent variable only. As a result, its ability to capture the coupled effects of complex ground conditions and varying operating states is limited, and it cannot reliably estimate thrust under such circumstances. Physical model tests are an important approach for revealing the coupling mechanisms between shield thrust and other tunnelling parameters. Nevertheless, when scaling prototype conditions down for laboratory testing, it is often difficult to simultaneously satisfy both geometric similarity and similarity in material mechanical properties. Zhu Hehua et al. [6] analyzed the relationship between shield thrust and torque at different burial depths through laboratory model tests in sandy soils, and Su et al. [7] conducted reduced-scale model tests to study the coupling behavior between shield thrust and soil chamber pressure. These studies are useful for revealing the mechanisms associated with one or a few influencing factors, but they find it difficult to account simultaneously for the combined effects of geological conditions, tunnel geometry and multiple tunnelling parameters and thus have limited capability for accurately predicting shield thrust under complex conditions.
With advances in shield construction data acquisition systems and digitalization, large volumes of measured data—including geological parameters, tunnel geometry and multiple excavation parameters—have been accumulated in engineering practice, providing a solid data basis for the data-driven prediction of shield thrust. Recent studies have used machine learning methods to develop predictive models for shield tunnelling parameters. Sun et al. [8] proposed a dynamic prediction framework based on multi-source heterogeneous data, using a random forest model to predict shield loads, thereby providing a paradigm for data-driven research on shield tunnelling parameters. Kong et al. [9] developed a random-forest-based prediction model for thrust and torque by fully integrating geological conditions and tunnelling parameters, and its prediction accuracy was significantly better than that of traditional empirical methods. Liu et al. [10] constructed characteristic indicators such as tunnelling specific energy and a field cutting depth index and used artificial neural networks and support vector regression, respectively, to predict key tunnelling parameters, thereby providing references for evaluating shield energy consumption and optimizing tunnelling parameter configurations. Abbasi et al. [11] investigated an EPB-TBM project on Mashhad Metro Line 3 in Iran and systematically compared the performance of multiple machine-learning models for predicting TBM performance/penetration rate, providing a useful reference for evaluating the engineering applicability of different algorithms in practice.
Most of the above machine-learning models treat the tunnelling data from each ring as an independent sample, implicitly assuming that there is no correlation between rings. However, shield tunnelling is inherently a dynamic process with strong temporal dependence, and the thrust state at the current ring is inevitably influenced by the “inertia” of parameters from previous rings. Ignoring the time-series characteristics of the tunnelling parameters may prevent the model from fully capturing the dynamic patterns governing thrust evolution, thereby degrading prediction accuracy and generalization performance.
To address this limitation, temporal deep learning models have gradually been introduced into the prediction of shield tunnelling parameters in recent years [12,13]. Long Short-Term Memory (LSTM) networks can selectively retain or discard historical information during training through gating mechanisms formed by the input, forget and output gates and are well suited to engineering time-series data with long-term dependencies. Existing studies have shown that time-series deep learning models represented by LSTM can effectively learn the nonlinear mapping from historical monitoring data to future parameter values, enabling the more accurate prediction of key quantities such as thrust. For example, Chen et al. [14] developed an LSTM model using short-term sequence data from the “ascending phase” of the tunnelling cycle to predict tunnelling parameters in the subsequent steady phase, thereby demonstrating the feasibility of the dynamic prediction of tunnelling parameters based on LSTM. Qiu et al. [15] constructed an LSTM model to accurately predict shield thrust and torque for deep-buried tunnel projects and achieved high prediction accuracy under complex surrounding rock conditions. Li et al. [16] further proposed a hybrid deep learning framework that combines Kriging interpolation, a channel-attention convolutional neural network and LSTM, effectively integrating geological and tunnelling features and significantly improving the accuracy of shield attitude prediction in complex ground conditions.
In summary, temporal deep learning models provide a powerful tool for the dynamic, data-driven prediction of shield tunnelling parameters, particularly shield thrust.
When using deep learning methods for thrust prediction, the neural network architecture and hyperparameter settings have a significant impact on model performance. The bidirectional long short-term memory network (BiLSTM) extends the conventional LSTM by introducing two recurrent structures that process the sequence in the forward and backward directions. This enables the model to exploit dependencies both from past to future and from future to past within the input sequence for feature extraction, thereby capturing more comprehensively the temporal evolution patterns of thrust and other parameters [17]. Hyperparameters of BiLSTM models—such as the number of neurons, learning rate and dropout rate—usually need to be specified before training, and their values directly affect the convergence speed and prediction accuracy of the network. However, it is difficult to determine an optimal hyperparameter combination purely based on experience. Manual trial-and-error or grid search is not only time-consuming but is also prone to becoming trapped in local optima, which limits model performance.
To overcome these problems, researchers have introduced intelligent optimization algorithms to optimize the hyperparameters of deep learning models [18]. Mahmoodzadeh et al. [19] combined an LSTM network with the grey wolf optimizer to predict TBM tunnelling performance, demonstrating that hybrid frameworks integrating temporal deep learning and intelligent optimization have strong potential for data-driven prediction in tunnel engineering. Genetic algorithms (GA), particle swarm optimization (PSO), the sparrow search algorithm (SSA) and the Hunger Games Search (HGS) algorithm all exhibit strong global search capabilities, do not rely on gradient information of the objective function and are well suited for hyperparameter search in complex, high-dimensional spaces [20]. Ye et al. [21] used PSO to optimize the number of neurons in LSTM and BPNN models, significantly improving the prediction accuracy of tunnelling speed. Gorgolis et al. [22] showed that the prediction accuracy of an LSTM model optimized by GA was significantly and consistently higher than that of an unoptimized baseline model. Zhang et al. [23] used an SSA-optimized LSTM model for real-time multi-region thrust prediction. Meanwhile, Wang et al. [24] combined HGS with LSTM and demonstrated that this approach can effectively improve model accuracy in shield tunnelling parameter prediction tasks. Hyperparameters have a substantial impact on the performance of temporal neural networks; therefore, selecting appropriate hyperparameters is crucial for achieving high prediction accuracy. Because optimization algorithms differ markedly in their search strategies, identifying the most suitable method for a given dataset or engineering problem is particularly important. Accordingly, this study compares different optimizers in terms of prediction accuracy, computational efficiency and stability during the hyperparameter tuning process, with the aim of determining an algorithm that delivers high accuracy, fast convergence and robust performance.
In summary, intelligent optimization algorithms offer clear advantages for hyperparameter optimization in deep learning models. However, most existing studies have focused on issues such as the tunnelling rate and cutterhead torque, and research on shield thrust prediction remains relatively limited. At the same time, many studies employ only a single optimization algorithm, and systematic comparisons of different optimization strategies in terms of prediction accuracy, convergence efficiency and result stability are still lacking.
Based on the above analysis, this study develops a BiLSTM-based time-series prediction model for shield thrust that simultaneously accounts for the effects of geological conditions, tunnel geometry and multiple tunnelling control parameters. Four intelligent optimization algorithms—GA, HGS, PSO and SSA—are employed to automatically tune key hyperparameters of the BiLSTM model. By comparing BiLSTM models optimized by different algorithms in terms of prediction accuracy, convergence speed and sensitivity to population size, the objectives of this study are twofold: (1) to establish a dynamic prediction framework for shield thrust that is applicable to complex ground conditions; and (2) to evaluate the characteristics of different intelligent optimization algorithms in the hyperparameter optimization of shield thrust prediction models.

2. Methodological Framework

An intelligent hybrid modeling framework for shield thrust prediction is constructed by jointly optimizing the number of neurons, learning rate and dropout rate of a BiLSTM model using GA, HGS, PSO and SSA.

2.1. Computational Principles of the BiLSTM Neural Network

BiLSTM is an extension of the traditional LSTM. LSTM is a special type of recurrent neural network (RNN) designed to address the vanishing and exploding gradient problems that occur in traditional RNNs when training on long sequences. As shown in Figure 1, LSTM effectively handles complex long-sequence dependencies by introducing gating mechanisms and a cell state to control the flow of information, thereby enabling selective forgetting, memorizing and outputting. The computation process is as follows:
(1) Forget gate: determines how much information from the previous cell state c t 1 should be retained:
f t = σ ( W f [ h t 1 , x t ] + b f )
(2) Input gate and candidate cell state: decides how much of the current input x t is written into the cell state while generating the candidate cell state c ˜ t :
i t = σ ( W i [ h t 1 , x t ] + b i )
c ˜ t = tanh ( W c [ h t 1 , x t ] + b c )
(3) Cell state update: combines the effects of the forget and input gates to update the cell state:
c t = f t c t 1 + i t c ˜ t
(4) Output gate: based on the updated cell state, computes the final hidden state h t :
o t = σ ( W o [ h t 1 , x t ] + b o )
h t = o t tanh ( c t )
In the above equations, x t denotes the input vector at time step t, and h t 1 is the hidden state at the previous time step.
The notation [ h t 1 , x t ] represents the concatenation of the two vectors. The variables f t , i t and o t are the activations of the forget, input and output gates, respectively; c t is the cell state at time step t, and c ˜ t is the candidate cell state. W f , W i , W c and W o are the corresponding weight matrices, and b f , b i , b c and b o are the corresponding bias vectors. σ denotes the Sigmoid activation function, tanh denotes the hyperbolic tangent activation function and denotes element-wise multiplication.
By combining two LSTMs that process the sequence in opposite directions, a BiLSTM is formed, as shown in Figure 2. In a BiLSTM, at each time step t, the forward layer captures the dependency of historical information from the beginning of the sequence up to t, while the backward layer captures the intra-sequence dependencies in the reverse direction (from later to earlier time steps) within the given input sequence. After concatenating the hidden states from both directions, h t integrates the forward and backward contextual information within the same historical window, thereby providing a richer temporal feature representation.

2.2. Hyperparameters of BiLSTM

In BiLSTM networks, hyperparameter settings have a significant impact on the convergence speed, representational capacity and generalization performance of the model. Among them, the number of neurons, the learning rate and the dropout rate are three of the most sensitive hyperparameters affecting model performance. Proper hyperparameter configuration helps improve the accuracy of shield thrust prediction, whereas inappropriate values may lead to underfitting, overfitting or unstable training.
In a BiLSTM, the number of neurons usually refers to the dimensionality of the hidden state in each LSTM layer, i.e., the number of units in that layer. Each neuron can be regarded as a feature extractor, and the number of neurons determines the capacity of the model, namely, its ability to learn complex patterns and retain information. With more neurons, richer features can be extracted. However, if there are too few neurons, the model may fail to capture complex relationships in the data; if there are too many neurons, the model is prone to overfitting and unnecessary computational cost.
The learning rate controls the step size of each parameter update and thus governs both the speed and stability of convergence. A learning rate that is too large may cause oscillation or even divergence of the training process, whereas a learning rate that is too small may result in slow convergence and increase the risk of getting trapped in local optima. An appropriately chosen learning rate helps the model approach a good solution efficiently while maintaining good generalization performance.
The dropout rate specifies the proportion of neurons that are randomly “dropped” during training. Dropout is an effective regularization technique whose main purpose is to prevent overfitting, i.e., the situation where the model fits the training data too well but performs poorly on unseen test data. When the dropout rate is too low, the regularization effect is weak and the model may still overfit. When the dropout rate is too high, the regularization effect is overly strong, which can destroy too much information and cause underfitting, making it difficult for the model to learn effective representations from the data.
In this study, the number of neurons, the learning rate and the dropout rate are selected as the hyperparameters to be optimized for the BiLSTM model. An intelligent optimization algorithm is employed to jointly optimize these hyperparameters in order to further improve the accuracy and stability of shield thrust prediction.

2.3. Intelligent Optimization Algorithm

To improve the prediction accuracy and generalization performance of the BiLSTM model and to overcome the limitations of manual, experience-based hyperparameter selection, this study employs intelligent optimization algorithms, namely, GA, HGS, PSO and SSA, to automatically tune three key hyperparameters of the BiLSTM model.
The fitness function is the key indicator for evaluating the quality of hyperparameter combinations. In this study, the root-mean-square error (RMSE) between the predicted and measured values is used as the fitness value of each individual [19,25,26]. Each candidate hyperparameter configuration is trained for 100 epochs on the training set. RMSE assigns a higher penalty to larger errors, which helps suppress peak deviations in safety-sensitive engineering parameter prediction tasks; moreover, it shares the same unit as thrust, facilitating engineering interpretation and comparisons across different models. A smaller RMSE indicates better predictive performance of the model under the corresponding hyperparameter set. Let θ i ( t ) denote the hyperparameter vector of the i -th individual in generation t ; then the fitness function is defined as
R M S E i ( t ) = 1 M j = 1 M y j y ^ j ( θ i ( t ) ) 2
M is the number of training samples, y j is the measured thrust of sample j and y ^ j ( θ i ( t ) ) is the predicted thrust for sample j under the hyperparameter vector θ i ( t ) .

2.3.1. Genetic Algorithm

GA is an intelligent optimization algorithm that mimics natural selection and genetic evolution. To be consistent with the other optimization algorithms used in this study, a real-coded GA is adopted, in which each candidate BiLSTM hyperparameter combination is directly encoded as a chromosome of real-valued genes. The evolution of GA is mainly driven by three basic operators: selection, crossover and mutation.
As shown in Figure 3, the selection operator constructs a breeding pool according to individual fitness; individuals with higher fitness are more likely to be retained and selected for reproduction, thereby simulating the natural selection principle of “survival of the fittest”. The crossover operator then selects pairs of parent individuals from the breeding pool and, with crossover probability P c , exchanges segments of their genes to generate offspring, achieving information recombination and expanding the search space. In this study, P c is set to 0.95. Finally, the genes of the offspring individuals are randomly mutated with mutation probability P m to generate new gene values, maintain population diversity and reduce the risk of premature convergence to a local optimum.
In the real-coded representation, the mutation of the j -th gene of the i -th individual can be expressed as
X i , j n e w = X i , j o l d + δ , if   rand < P m , X i , j o l d , or
where X i , j o l d and X i , j n e w denote the gene values before and after mutation, respectively; P m is the mutation probability, which is set to 0.025 in this study; δ is a random perturbation generated according to the specified distribution; and rand is a random number in the interval [0, 1]. By repeatedly applying selection, crossover and mutation operations over multiple generations, the GA population gradually evolves towards regions with lower fitness values in the hyperparameter search space, ultimately yielding an optimal hyperparameter configuration for the GA–BiLSTM shield thrust prediction model.

2.3.2. Hunger Games Search Algorithm

HGS is an intelligent optimization algorithm inspired by animal foraging behavior. By modeling individuals’ hunger levels and their foraging strategies under different game modes, it seeks to balance global exploration and local exploitation in the search space. By introducing hunger-driven adaptive weights and stochastic perturbation terms, HGS can, to some extent, mitigate premature convergence and enhance the ability to escape local optima; however, its convergence performance is still influenced by factors such as population size, dataset characteristics and specific engineering scenarios.
As shown in Figure 4, during each iteration, HGS first ranks the population according to individual fitness and then computes the hunger level and corresponding weight for each individual. Hunger is usually inversely proportional to fitness: individuals with lower fitness have higher hunger and therefore obtain stronger search momentum during position updates. Denoting the global best position in generation t by θ b e s t t and the random perturbation term for individual i by Δ i t , the position of individual i in generation t + 1 can be written as
θ i t + 1 = θ i t + E i t ( θ b e s t t θ i t ) + Δ i t
where E i t is a weight coefficient determined by the hunger level and control parameters, and Δ i t introduces randomness and diversity. By adopting different game modes, HGS enhances global exploration in the early stage of the search and gradually strengthens local exploitation in the later stage, thereby guiding the population to converge towards regions with higher fitness.

2.3.3. Particle Swarm Optimization Algorithm

PSO is an intelligent optimization algorithm that simulates the foraging behavior of bird flocks. In PSO, each candidate hyperparameter combination is represented as a particle position vector X i in the search space, and the particles are iteratively updated according to their current velocity, their own best historical position and the global best position of the swarm so as to approach the optimal solution. As shown in Figure 5, in generation t , the updates of the velocity V i t and position X i t of particle i can be expressed as
V i t + 1 = w V i t + c 1 r 1 ( P b e s t , i t X i t ) + c 2 r 2 ( X b e s t t X i t )
X i t + 1 = X i t + V i t + 1
w t = w u t ( w u w d ) n m a x
where P b e s t , i t is the personal best position found so far by particle i ; X b e s t t is the global best position of the current swarm; c 1 and c 2 are the cognitive and social learning factors, respectively, and are both set to 2 in this study; r 1 and r 2 are random numbers uniformly distributed in [0, 1], used to increase the randomness and diversity of the search; w t is the inertia weight, which trades off global exploration and local exploitation; w u and w d denote the maximum and minimum inertia weights, set to 0.9 and 0.4 in this study, respectively; and n m a x is the maximum number of iterations. By iteratively applying the above update rules, the particle swarm gradually converges toward regions with higher fitness in the hyperparameter search space, yielding an improved hyperparameter configuration for the BiLSTM model.

2.3.4. Sparrow Search Algorithm

SSA is an intelligent optimization algorithm that mimics the foraging and anti-predation behaviors of sparrows. Individuals in the population are divided into producers and scroungers according to their functional roles, and a small proportion of them act as scouts to detect danger signals. Producers are mainly responsible for searching for food sources over a wide area; scroungers follow the producers to forage in the surrounding area; and scouts guide the population out of the current region when danger is detected, thereby reducing the risk of falling into local optima.
As shown in Figure 6, in generation t , the position vector of sparrow i is denoted by X i t . Depending on the role of the individual and the current danger level, the update of X i t can be written as
X i t + 1 = X i t exp i α T max , if   producer   and   R 2 < S T X b e s t t + | X i t X b e s t t | A + L , if   scrounger X b e s t t + β | X i t X b e s t t | , if   scout
where X b e s t t is the global best position at generation t ; T max is the maximum number of iterations; α is a control parameter; R 2 is a random number uniformly distributed on [0, 1] that characterizes the danger level; S T is the safety threshold; A + is a random matrix generated from a truncated normal distribution; L is a random vector uniformly distributed on [−1, 1]; and β is a constant that controls the step size. Through this mechanism, producers are responsible for guiding global exploration, scroungers perform local exploitation around high-quality individuals and scouts drive the population away from dangerous or stagnant regions, thus achieving a balance between exploration and exploitation.

2.4. Prediction Method Framework

Using measured tunnel geometry, geomechanical properties, and shield tunnelling parameters from the Guanzhuang–Yongshun section of Beijing Metro Line 22 as input variables, next-ring TH was selected as the prediction target. A total of 1500 ring-wise samples were collected. To preserve the integrity of time-series prediction, the training/testing split strictly followed the chronological order of ring numbers: the first 1200 sequences were used for training and the remaining 300 sequences were used for testing. No random shuffling was performed to avoid disrupting temporal dependencies and artificially inflating performance. To mitigate the influence of differing units and magnitudes among physical variables on network training, feature normalization was carried out using the linear transformation in Equation (14), mapping the data to the interval [−1, 1]. The normalization parameters were determined consistently over the full dataset to ensure scale consistency when comparing different algorithms. After verification, the minimum and maximum values of each feature in the test set fall within the range of the training set; therefore, the min/max values computed from the full dataset are identical to those computed using only the training set in Equation (14). Hence, no information leakage is introduced by the normalization procedure.
P n = 2 ( X X min ) ( X max X min ) 1
In the formula, P n are the normalized values, and X max and X min are the maximum and minimum values of the measured data.
The time step was set to 20. That is, at any time t, the model uses only the parameters from the previous 20 excavated rings (t – 19, …, t) as inputs to predict the thrust of the next ring (t + 1). Therefore, during both training and testing, the model does not use any observations from future ring numbers. The backward branch of the BiLSTM performs feature extraction only within this historical window to enhance sequence representation.
All experiments were conducted on Ubuntu 22.04, using Python 3.12 and PyTorch 2.8.0. The computing resources included an NVIDIA RTX 5090 GPU and an Intel(R) Xeon(R) Gold 6459C CPU.
The number of neurons, learning rate and dropout rate were selected as the hyperparameters to be optimized, with search ranges set to the number of neurons [20, 100], learning rate [0.001, 0.05] and dropout rate [0, 0.7]. Four intelligent optimization algorithms—GA, PSO, HGS and SSA—were adopted, and three population sizes (10, 30 and 50) were considered for the search. The maximum number of iterations for GA, PSO, HGS and SSA was set to 100, and the stopping criterion was reaching the maximum iteration number. Since both BiLSTM training and intelligent optimization involve stochastic mechanisms such as random initialization, the fitness values from a single run may exhibit some variability. Therefore, each combination of algorithm and population size was executed independently 10 times. The root mean square error (RMSE) between predicted and measured values was used as the fitness function, and the best hyperparameter set from each run was recorded. Based on these results, four hybrid prediction models—GA-BiLSTM, HGS-BiLSTM, PSO-BiLSTM and SSA-BiLSTM—were constructed, as shown in Figure 7.

3. Database Construction

Using data from the shield tunnelling section of the Beijing Metro Line 22 project from Guanzhuang to Yongshun, this study selects 11 characteristic variables from three aspects—tunnel geometry, geomechanical conditions and shield tunnelling parameters—to construct a multi-source database for shield thrust prediction. Specifically, one tunnel geometry parameter is considered, namely, the tunnel burial depth ( h ); five geomechanical parameters are considered, namely, cohesion ( c ), equivalent internal friction angle ( φ ), compression modulus ( C M ), permeability coefficient ( P C ) and standard penetration blow count ( N ); and five shield tunnelling parameters are considered, namely, excavation speed ( E S ), cutterhead torque ( C H T ), soil chamber pressure ( C P ), synchronous grouting pressure ( G P ) and synchronous grouting volume ( G V ).
The burial depth of each ring of the shield tunnel is obtained from the geological investigation report, while the five shield tunnelling parameters are obtained from on-site construction records. To account for the influence of geomechanical conditions on model predictions, the geomechanical parameters of all strata from the ground surface to the tunnel crown within the excavation section of each ring are computed by weighted averaging, and the calculation formula is given as follows:
X = i = 1 n t i h h i h X i
In the formula, X i is the geomechanical parameter; n is the number of strata; t i is the thickness of the first layer of soil; h i is burial depth of the i -th layer of soil; and h is tunnel burial depth.
To ensure the reliability and temporal consistency of the modelling data, the raw ring-wise records were preprocessed before model development. First, the construction logs were aligned by ring number to ensure a one-to-one correspondence among geometric parameters, geological parameters and tunnelling parameters for the same ring. For occasional missing entries, adjacent-ring interpolation was used when the missing ratio was small, whereas records that could not be reliably recovered were removed. Second, considering potential abnormal records during shield tunnelling—such as abrupt changes caused by shutdowns, operating-mode switching or sensor fluctuations—statistical screening was performed using the 3σ rule and further checked against engineering-reasonable ranges. Isolated outliers were replaced with smoothed values to maintain sequence continuity, while consecutive abnormal segments were discarded to avoid biasing the temporal learning process. After these procedures, continuous and valid samples were obtained for model training and testing.
After the above processing, 1500 samples were obtained. Each sample contains 12 variables, including 11 input variables and 1 output variable. The maximum, minimum, mean and median values of each parameter are summarized in Table 1. For the ground parameters, the burial depth h ranges from 17.7 to 33.0 m, with a mean of 25.3 m and a median of 24.7 m; cohesion c ranges from 4.8 to 21.5 kPa, with a mean of 12.7 kPa and a median of 12.0 kPa; and the internal friction angle φ ranges from 14.6° to 24.9°, with a mean of 20.2° and a median of 20.7°.
For the shield tunnelling parameters, the excavation speed E S ranges from 40 to 97 mm/min, with a mean of 72.7 mm/min and a median of 73 mm/min. The synchronous grouting volume G V ranges from 10.0 × 103 to 16.5 × 103 L, with a mean of 12.1 × 103 L and a median of 12.1 × 103 L. The output variable, total shield thrust T H , ranges from 22.0 × 103 to 49.0 × 103 kN, with a mean of 32.9 × 103 kN and a median of 32.0 × 103 kN.
The distributions of the 11 input variables and the 1 output variable are shown in Figure 8. It can be observed that each parameter generally exhibits a unimodal distribution: samples are mainly concentrated around their respective mean values, with a small number of samples located in the tails on both sides. This indicates that the data not only cover the typical operating conditions of the project site but also include a certain proportion of extreme conditions, providing a rich sample basis for subsequent training and generalization of the BiLSTM model.
To avoid qualitative judgments based solely on the distribution shape, the tail samples were further quantified using the quantiles of the ground-truth thrust in the test set. Specifically, 22 samples fall in the lower tail (TH ≤ P5) and 21 samples fall in the upper tail (TH ≥ P95), accounting for 7.3% and 7.0% of the test set, respectively. This indicates that the test set covers not only typical operating conditions but also a certain number of extreme high/low thrust cases.

4. Analysis of Shield Thrust Prediction Results

Based on the four hybrid models GA–BiLSTM, HGS–BiLSTM, PSO–BiLSTM and SSA–BiLSTM, the performance of shield thrust prediction is comprehensively analyzed in terms of the influence of population size on convergence characteristics and computational efficiency, as well as prediction accuracy under evaluation indices such as RMSE, MAE, MAPE and R2.

4.1. Sensitivity Analysis of Computational Efficiency with Respect to Population Size

To analyze the impact of population size on optimization performance and computational efficiency, three population sizes, Pop = 10, 30 and 50, are selected for comparative experiments using the four algorithms GA, HGS, PSO and SSA. Figure 9 presents the average number of iterations to convergence and the computation time under different population sizes, and the corresponding numerical values are listed in Table 2.
In terms of sensitivity to population size, the average number of iterations of GA and PSO varies relatively smoothly as Pop changes, whereas HGS and SSA are more sensitive to population size. When Pop = 10, the numbers of iterations for HGS and SSA are 44.6 and 42.6, respectively, which are significantly lower than those for Pop = 30 and 50. However, their average fitness values are 1961.73 kN and 1963.22 kN, both higher than 1944.10 kN and 1933.98 kN for Pop = 30 and 1941.07 kN and 1934.22 kN for Pop = 50. This indicates that, although convergence is faster with a small population, the quality of the obtained solutions is poorer and the algorithms tend to suffer from pronounced premature convergence. As the population size increases to 30, the average fitness of HGS and SSA decreases markedly and stabilizes around 1944.10 kN and 1933.98 kN, respectively, thereby avoiding premature convergence while ensuring a more thorough search.
In terms of computational time, the runtime of all four algorithms increases significantly as the population size grows. When Pop increases from 10 to 50, the computation times of GA, HGS, PSO and SSA rise from 9.02, 6.10, 9.22 and 11.72 min to 58.86, 38.00, 43.72 and 67.88 min, respectively, i.e., approximately five to six times the runtime at Pop = 10.
Considering both optimization performance and computational efficiency, the average fitness differences among the four algorithms are small for both Pop = 30 and Pop = 50, whereas Pop = 30 significantly reduces computation time compared with Pop = 50. Therefore, in the subsequent performance comparison experiments in this study, Pop = 30 is uniformly adopted as the benchmark population size, which ensures reliable convergence of the algorithms while also satisfying the time requirements of practical engineering applications.
Table 3 reports the test RMSE and corresponding rankings of the optimal configurations obtained by each algorithm under Pop = 10/30/50. It can be observed that the relative ranking of the four hybrid models remains consistent across different population sizes, indicating that the conclusions regarding the comparative performance of the algorithms are robust with respect to Pop. In terms of accuracy, the overall difference between Pop = 10 and Pop = 30 is small, whereas Pop = 50 further reduces the error for some algorithms. However, when the computational time and convergence efficiency shown in Figure 9 and Table 2 are jointly considered, the time cost of Pop = 50 increases substantially. Meanwhile, although the small population size Pop = 10 can achieve comparable accuracy, it is more sensitive to random initialization and entails a higher risk of premature convergence, making it unsuitable as a unified benchmark for fair comparison. Therefore, Pop = 30 is selected as the uniform setting in subsequent analyses to strike a balance among accuracy, efficiency and stability.

4.2. Comparison of Thrust Prediction Performance

To systematically evaluate the predictive performance of the four hybrid models, four metrics are computed on the same test set: root-mean-square error (RMSE), mean absolute error (MAE), mean absolute percentage error (MAPE) and the coefficient of determination (R2). The results are summarized in Table 4.
As shown in Table 4, the RMSE values of the four models are approximately 1330–1400 kN, the MAE values lie between 1030 kN and 1110 kN, all MAPE values are below 3.5% and all R2 values are around 0.70. This indicates that, after the introduction of intelligent optimization algorithms, the BiLSTM-based models as a whole can adequately capture the nonlinear relationship between shield thrust and its influencing factors. Nevertheless, there are noticeable differences among the optimization algorithms. Among the four models, HGS–BiLSTM achieves the best performance on all four metrics, with RMSE = 1332 kN, MAE = 1032 kN, MAPE = 3.17% and R2 = 0.726, indicating that it outperforms the other three models in terms of both error level and goodness of fit. The errors of PSO–BiLSTM are slightly higher than those of HGS–BiLSTM but lower than those of GA–BiLSTM. SSA–BiLSTM performs the worst among the four models, with RMSE = 1400 kN, MAE = 1106 kN, MAPE = 3.38% and R2 = 0.697, indicating relatively weak overall predictive performance.
Meanwhile, to examine the reliability of the models under extreme thrust conditions, error metrics were computed separately on the tail subsamples TH ≤ P5 and TH ≥ P95, as reported in Table 5. As shown in Table 5, the errors on the extreme-thrust subsamples TH ≤ P5 or TH ≥ P95 are generally higher than the overall statistics in Table 4, indicating that prediction becomes more challenging in tail conditions. For TH ≥ P95, the model errors diverge markedly: PSO–BiLSTM achieves the lowest RMSE of 931.98 kN and MAE of 639.84 kN in this range, followed by HGS–BiLSTM. In contrast, for TH ≤ P5, all models exhibit relatively large errors, and PSO–BiLSTM shows the highest error, suggesting comparatively weaker robustness under low-thrust conditions. Overall, the tail-subsample analysis confirms performance differences among the models under extreme operating conditions and provides quantitative evidence to support model selection for risk-critical engineering scenarios.
To further elucidate the error distribution patterns across different thrust levels, Figure 10 presents the variation of MAE after binning the test set into quantile-based intervals. As shown, the error exhibits a “larger at both ends and smaller in the middle” trend: MAE increases noticeably in the low- and high-thrust ranges while remaining relatively low across the mid-range (P25–P95). This indicates that the models provide more stable predictions under typical operating conditions but are more prone to larger deviations under extreme thrust conditions; therefore, engineering applicability should be assessed comprehensively in conjunction with tail-focused metrics.
For a more intuitive comparison of the predictive performance of each model, Figure 11 presents the fitted curves and error distributions of the four hybrid models on the test set. It can be observed that the prediction curve of HGS–BiLSTM matches the measured values more closely and exhibits smaller residual fluctuations, whereas SSA–BiLSTM shows noticeable overestimation or underestimation in some peak segments. As indicated by Table 4 and Figure 11, HGS–BiLSTM achieves the best overall performance on the full-sample metrics. Meanwhile, Table 5 and Figure 10 show that the relative advantages of different models vary in the extreme tail regions. Therefore, HGS–BiLSTM is selected as the overall preferred model, and it is recommended that tail-focused error metrics be considered under extreme operating conditions to support decision-making.

5. Conclusions

Taking the shield tunneling section of the Beijing Metro Line 22 project from Guanzhuang to Yongshun as the engineering background, this study develops four hybrid models, namely, GA–BiLSTM, HGS–BiLSTM, PSO–BiLSTM and SSA–BiLSTM, for shield thrust prediction and systematically compares their performance. The main conclusions are as follows:
(1) The BiLSTM sequence model constructed using tunnel geometric parameters, geomechanical parameters and shield tunnelling parameters can effectively capture the nonlinearity and temporal dependence of shield thrust. After incorporating the intelligent optimization algorithms GA, HGS, PSO and SSA, the four hybrid models all achieved accurate one-step-ahead prediction on the Guanzhuang–Yongshun dataset of Beijing Metro Line 22, with test-set MAPE values below 3.5%. Although the errors increase under extreme tail operating conditions, they remain generally acceptable, indicating that the models exhibit a certain degree of reliability in extreme thrust ranges. Overall, the proposed framework demonstrates effectiveness and engineering applicability for the studied section under the adopted time-series split and modeling settings; however, its transferability still needs further validation on additional sections and projects.
(2) Population size has a significant impact on algorithm convergence and computational efficiency. For the dataset used in this study, HGS and SSA with a small population size Pop = 10 are prone to premature convergence, whereas an excessively large population Pop = 50 leads to a multi-fold increase in computational time. By jointly considering the average fitness, iteration number and computational time, Pop = 30 can effectively mitigate premature convergence while keeping the computational cost under control and thus can be regarded as a reasonable population-size setting for similar engineering applications.
(3) For the dataset used in this study, under the condition of uniformly adopting Pop = 30, among the four hybrid models, HGS–BiLSTM exhibits the best overall performance, with the smallest RMSE, MAE and MAPE and the largest R2. It can therefore be recommended as a preferred model for real-time prediction of shield thrust and construction decision support under complex geological conditions.
It should be noted that the conclusions of this study are based on a single project section and a fixed time-series split; further validation of the model’s generalizability will be conducted using data from multiple sections or multiple projects.

Author Contributions

Conceptualization, L.Y. and F.K.; methodology, L.Y. and F.K.; software, L.Y. and W.Y.; validation, L.Y. and F.K.; formal analysis, L.Y.; investigation, L.Y.; resources, W.Y.; data curation, W.Y.; writing—original draft preparation, L.Y.; writing—review and editing, F.K. and J.W.; visualization, L.Y.; supervision, F.K. and J.W.; project administration, F.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets generated and analyzed during the current study, as well as the source code used for model development and evaluation, are available from the corresponding author on reasonable request, due to commercial confidentiality restrictions.

Acknowledgments

The authors would like to thank the technical staff of Beijing Construction Engineering Group Co., Ltd. for their assistance with field data collection and TBM operation records.

Conflicts of Interest

Author Wei Yin was employed by the company Beijing Construction Engineering Group Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Editorial Department of China Journal of Highway and Transport. Review on China’s Traffic Tunnel Engineering Research: 2022. China J. Highw. Transp. 2022, 35, 1–40. [Google Scholar] [CrossRef]
  2. Hong, K. State of Art and Prospect of Tunnels and Underground Works in China. Tunn. Constr. 2015, 35, 95–107. [Google Scholar] [CrossRef]
  3. Hong, K. Development and Prospects of Tunnels and Underground Works in China in Recent Two Years. Tunn. Constr. 2017, 37, 123–134. [Google Scholar] [CrossRef]
  4. Zhou, C.; Ding, L.; Zhou, Y.; Skibniewski, M.J. Visibility graph analysis on time series of shield tunneling parameters based on complex network theory. Tunn. Undergr. Space Technol. 2019, 89, 10–24. [Google Scholar] [CrossRef]
  5. Krause, H. Geologische Erfahrungen beim Einsatz von Tunnelvortriebsmaschinen in Baden-Württemberg. In Latest Findings in the Construction of Underground Excavations—Rock Foundations; Springer: Berlin/Heidelberg, Germany, 1976; pp. 49–60. [Google Scholar]
  6. Zhu, H.-H.; Xu, Q.-W.; Liao, S.-M.; Yu, N.; Fu, D.-M.; Wu, X.-S. Experimental study on thrust force during tunnel excavation with earth pressure balance shield machine. Rock Soil Mech. 2007, 28, 1587–1594. [Google Scholar]
  7. Su, J.; Gong, G.; Yang, H. Total thrust calculation and test research for soil pressure balancing shield tunneling. Constr. Mach. Equip. 2008, 39, 13–15. [Google Scholar]
  8. Sun, W.; Shi, M.; Zhang, C.; Zhao, J.; Song, X. Dynamic load prediction of tunnel boring machine (TBM) based on heterogeneous in-situ data. Autom. Constr. 2018, 92, 23–34. [Google Scholar] [CrossRef]
  9. Kong, X.; Ling, X.; Tang, L.; Tang, W.; Zhang, Y. Random forest-based predictors for driving forces of earth pressure balance (EPB) shield tunnel boring machine (TBM). Tunn. Undergr. Space Technol. 2022, 122, 104373. [Google Scholar] [CrossRef]
  10. Liu, K.-Q.; Du, D.-C.; Zhao, W.; Ding, W.-T. Machine Learning-Based Prediction and Optimization of Slurry Shield’s Key Tunneling Parameters. J. Northeast. Univ. Nat. Sci. 2023, 44, 1621–1630. [Google Scholar] [CrossRef]
  11. Abbasi, M.; Namadchi, A.H.; Abbasi, M.; Abbasi, M. Evaluation of machine learning algorithms in tunnel boring machine applications: A case study in Mashhad metro line 3. Int. J. Geo-Eng. 2024, 15, 28. [Google Scholar] [CrossRef]
  12. Lu, D.; Ma, Y.; Kong, F.; Guo, C.; Miao, J.; Du, X. Support vector regression with heuristic optimization algorithms for predicting the ground surface displacement induced by EPB shield tunneling. Gondwana Res. 2023, 123, 3–15. [Google Scholar] [CrossRef]
  13. Kong, F.; Lu, D.; Ma, Y.; Li, J.; Tian, T. Analysis and Intelligent Prediction for Displacement of Stratum and Tunnel Lining by Shield Tunnel Excavation in Complex Geological Conditions: A Case Study. IEEE Trans. Intell. Transp. Syst. 2022, 23, 22206–22216. [Google Scholar] [CrossRef]
  14. Chen, H.; Xiao, C.; Yao, Z.; Jiang, H.; Zhang, T.; Guan, Y. Prediction of TBM Tunneling Parameters through an LSTM Neural Network. In Proceedings of the 2019 IEEE International Conference on Robotics and Biomimetics (ROBIO), Dali, China, 6–8 December 2019; pp. 702–707. [Google Scholar] [CrossRef]
  15. Qiu, D.; Fu, K.; Xue, Y.; Li, Z.; Li, G.; Kong, F. LSTM time-series prediction model for TBM tunneling parameters of deep-buried tunnels and application research. J. Cent. South Univ. Sci. Technol. 2021, 52, 2646–2660. [Google Scholar] [CrossRef]
  16. Li, P.N.; Liu, X.; Dai, Z.Y.; Fan, J.; Kou, X.; Zhai, Y. Dynamic prediction for attitude and position in shield tunneling based on hybrid deep learning method. J. Donghua Univ. Nat. Sci. 2024, 50, 145–152. [Google Scholar] [CrossRef]
  17. Lu, D.; Liu, Y.; Kong, F.; He, X.; Zhou, A.; Du, X. A novel Bi-LSTM method fusing current and historical data for tunnelling parameters of shield tunnel. Transp. Geotech. 2024, 49, 101402. [Google Scholar] [CrossRef]
  18. Kong, F.C.; Liu, Y.H.; Liu, S.J.; Lu, D.C.; Du, X.L. Intelligent prediction method for cutterhead torque of shield tunnel based on intelligent optimization algorithm and Bi-LSTM neural network. Sci. Sin. Technol. 2025, 55, 170–186. (In Chinese) [Google Scholar] [CrossRef]
  19. Mahmoodzadeh, A.; Nejati, H.R.; Mohammadi, M.; Ibrahim, H.H.; Rashidi, S.; Rashid, T.A. Forecasting tunnel boring machine penetration rate using LSTM deep neural network optimized by grey wolf optimization algorithm. Expert Syst. Appl. 2022, 209, 118303. [Google Scholar] [CrossRef]
  20. Armaghani, D.J.; Mohamad, E.T.; Narayanasamy, M.S.; Narita, N.; Yagiz, S. Development of hybrid intelligent models for predicting TBM penetration rate in hard rock condition. Tunn. Undergr. Space Technol. 2017, 63, 29–43. [Google Scholar] [CrossRef]
  21. Ye, X.-W.; Wei, Y.-J.; Zhang, X.-L.; Ding, Y.; Chen, Y.-M. Confining pressure forecasting of shield tunnel lining during construction based on LSTM–PSO models combined with the multi-output recursive strategy. Tunn. Undergr. Space Technol. 2024, 143, 105499. [Google Scholar] [CrossRef]
  22. Gorgolis, N.; Hatzilygeroudis, I.; Istenes, Z.; Gyenne, L. Hyperparameter Optimization of LSTM Network Models through Genetic Algorithm. In Proceedings of the 2019 10th International Conference on Information, Intelligence, Systems and Applications (IISA), Patras, Greece, 15–17 July 2019; pp. 1–4. [Google Scholar]
  23. Zhang, W.; Liu, X.; Zhang, L.; Wang, Y. Intelligent real-time prediction of multi-region thrust of EPB shield machine based on SSA-LSTM. Eng. Res. Express 2023, 5, 035013. [Google Scholar] [CrossRef]
  24. Wang, C.; Hua, D.-R.; Peng, J.; Huang, L.; Chen, Y.-Y. Prediction of Energy Consumption in Shipyard’s Production Workshops by Integrating EEMD and HGS-LSTM. Ship Ocean Eng. 2025, 54, 115–120,126. [Google Scholar] [CrossRef]
  25. Ulum, D.S.N.; Girsang, A.S. Hyperparameter Optimization of Long-Short Term Memory using Symbiotic Organism Search for Stock Prediction. Int. J. Innov. Res. Sci. Stud. 2022, 5, 121–133. [Google Scholar] [CrossRef]
  26. Suryawanshi, P.; Gaikwad, S.; Kumar, A.; Patlolla, A.; Jayakumar, S. A Hybrid Bayesian-Genetic Algorithm Based Hyperparameter Optimization of an LSTM Network for Demand Forecasting of Retail Products. In Proceedings of the 15th International Joint Conference on Computational Intelligence (IJCCI 2023)—ECTA, Rome, Italy, 13–15 November 2023; SciTePress: Setúbal, Portugal, 2023; pp. 230–237. [Google Scholar] [CrossRef]
Figure 1. Schematic diagram of the internal structure of an LSTM unit.
Figure 1. Schematic diagram of the internal structure of an LSTM unit.
Applsci 16 00325 g001
Figure 2. Schematic diagram of the BiLSTM network structure.
Figure 2. Schematic diagram of the BiLSTM network structure.
Applsci 16 00325 g002
Figure 3. Flowchart of the GA algorithm.
Figure 3. Flowchart of the GA algorithm.
Applsci 16 00325 g003
Figure 4. Flowchart of HGS algorithm.
Figure 4. Flowchart of HGS algorithm.
Applsci 16 00325 g004
Figure 5. Flowchart of the PSO algorithm.
Figure 5. Flowchart of the PSO algorithm.
Applsci 16 00325 g005
Figure 6. Flowchart of the SSA algorithm.
Figure 6. Flowchart of the SSA algorithm.
Applsci 16 00325 g006
Figure 7. Prediction results of the BiLSTM model optimized by different algorithms.
Figure 7. Prediction results of the BiLSTM model optimized by different algorithms.
Applsci 16 00325 g007
Figure 8. Distributions of key input variables and TH.
Figure 8. Distributions of key input variables and TH.
Applsci 16 00325 g008
Figure 9. Comparison of convergence time and number of iterations among the four algorithms.
Figure 9. Comparison of convergence time and number of iterations among the four algorithms.
Applsci 16 00325 g009
Figure 10. Comparison of MAE across thrust quantile intervals in the test set (Pop = 30).
Figure 10. Comparison of MAE across thrust quantile intervals in the test set (Pop = 30).
Applsci 16 00325 g010
Figure 11. Fitting curves and residual distribution of each model. (a) GA-BiLSTM; (b) HGS-BiLSTM; (c) PSO-BiLSTM; (d) SSA-BiLSTM.
Figure 11. Fitting curves and residual distribution of each model. (a) GA-BiLSTM; (b) HGS-BiLSTM; (c) PSO-BiLSTM; (d) SSA-BiLSTM.
Applsci 16 00325 g011
Table 1. Table of maximum, minimum, average and median values of 12 parameters.
Table 1. Table of maximum, minimum, average and median values of 12 parameters.
MaxMinMeanMedian
h (m)33.017.725.324.7
c (kPa)21.54.812.712.0
Phi (°)24.914.620.220.7
CM (MPa)11.15.88.28.1
PC (×10−3 m/d)2.660.501.311.17
N32.310.218.717.8
ES (mm/min)974072.773
CHT (×103 kN·m)12.04.07.17.0
CP (bar)2.61.11.81.8
GP (MPa)0.630.220.380.40
GV (×103 L)16.510.012.112.1
TH (×103 kN)49.022.032.932.0
Table 2. Statistics of convergence iterations, computation time and average fitness under different algorithms and population sizes.
Table 2. Statistics of convergence iterations, computation time and average fitness under different algorithms and population sizes.
AlgorithmPopulation SizeMean IterationsTime (min)Average Fitness (kN)
GA1066.09.021966.29
3069.031.311949.86
5068.558.861938.23
HGS1044.66.101961.73
3059.823.701944.10
5054.738.001941.07
PSO1065.99.221944.87
3068.928.541928.16
5058.543.721922.22
SSA1042.611.721963.22
3062.748.171933.98
5052.667.881934.22
Table 3. Comparison of test RMSE and rankings of the optimal configurations for each algorithm under different population sizes.
Table 3. Comparison of test RMSE and rankings of the optimal configurations for each algorithm under different population sizes.
AlgoPop = 10 RMSERankPop = 30 RMSERankPop = 50 RMSERank
GA136831368313293
HGS135811332112971
PSO136621362213232
SSA141541400414134
Table 4. Performance Metrics of the Four Models.
Table 4. Performance Metrics of the Four Models.
MethodsRMSE (kN)MAE (kN)MAPE (%)R2
GA-BiLSTM1368 1063 3.270.711
HGSBiLSTM1332 1032 3.170.726
PSO-BiLSTM1362 1036 3.230.714
SSA-BiLSTM1400 1106 3.380.697
Table 5. Comparison of error metrics on extreme tail-thrust subsamples with Pop = 30.
Table 5. Comparison of error metrics on extreme tail-thrust subsamples with Pop = 30.
TH_binAlgonRMSEMAE
≥P95GA211516.281185.04
≥P95HGS211239.63886.41
≥P95PSO21931.98639.84
≥P95SSA211365.891014.45
≤P5GA222371.592105.17
≤P5HGS222241.351883.55
≤P5PSO222573.742302.73
≤P5SSA222059.421653.38
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yao, L.; Yin, W.; Kong, F.; Wang, J. Shield Thrust Time-Series Prediction Based on BiLSTM with Intelligent Hyperparameter Optimization. Appl. Sci. 2026, 16, 325. https://doi.org/10.3390/app16010325

AMA Style

Yao L, Yin W, Kong F, Wang J. Shield Thrust Time-Series Prediction Based on BiLSTM with Intelligent Hyperparameter Optimization. Applied Sciences. 2026; 16(1):325. https://doi.org/10.3390/app16010325

Chicago/Turabian Style

Yao, Lingbin, Wei Yin, Fanchao Kong, and Junqi Wang. 2026. "Shield Thrust Time-Series Prediction Based on BiLSTM with Intelligent Hyperparameter Optimization" Applied Sciences 16, no. 1: 325. https://doi.org/10.3390/app16010325

APA Style

Yao, L., Yin, W., Kong, F., & Wang, J. (2026). Shield Thrust Time-Series Prediction Based on BiLSTM with Intelligent Hyperparameter Optimization. Applied Sciences, 16(1), 325. https://doi.org/10.3390/app16010325

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop