1. Introduction
In contemporary transportation, energy, and infrastructure projects, tunnels are extensively employed to ensure continuity of intercity connections, support water supply and wastewater management systems, and enable the safe transmission of energy. As these structures are often constructed under complex geotechnical conditions, they require a high level of technical expertise, meticulous planning, and advanced safety measures throughout the design, construction, and operational phases. Tunnel boring machines (TBMs) represent one of the most tangible field-level manifestations of this multidisciplinary framework. Accurate prediction of TBM performance relies on the integrated knowledge of multiple disciplines, including civil engineering, geology, mechanical engineering, data science, and project management. Numerous parameters—such as the rate of penetration (ROP), advance rate (AR), and ground–support interactions—directly influence the efficiency and safety of TBM operations. Reliable estimation of these parameters is critical not only for occupational safety but also for effective cost and time management. Consequently, the success of a TBM project necessitates strong teamwork and effective communication among geologists, geotechnical engineers, mechanical engineers, data analysts, and field personnel, while clearly highlighting the importance of interdisciplinary approaches and team coordination in engineering education and practice.
Studies on the prediction of TBM penetration rate initially focused on simple empirical models based on rock hardness [
1]. Subsequently, investigations into penetration behavior in sedimentary rock environments [
2] and statistical regression models [
3,
4] contributed significantly to understanding the fundamental relationships between rock properties and ROP. Advanced regression analyses yielded high correlation coefficients (r = 0.82), leading to the development of empirical equations applicable to practical engineering problems [
3]. However, as these early approaches largely relied on linear assumptions, they proved insufficient in fully capturing complex geotechnical interactions and machine–ground behavior.
To achieve a more realistic assessment of TBM performance, heuristic optimization algorithms have been increasingly adopted in subsequent studies. A pioneering contribution in this area is the Particle Swarm Optimization (PSO)—based model proposed by Yagiz and Karahan [
5], which aimed to minimize the discrepancy between measured and predicted ROP values. Further studies expanded methodological diversity by incorporating heuristic algorithms such as Harmony Search (HS), Differential Evolution (DE), and the Grey Wolf Optimizer (GWO) [
6]. Similarly, Khoshzaher et al. [
7] demonstrated that TBM penetration rate is influenced by both machine-related and geological factors, showing that the Firefly algorithm outperformed PSO in terms of optimization efficiency.
In complex geotechnical and mechanical systems such as TBMs, it is extremely challenging to represent all governing relationships comprehensively within a single analytical expression. This limitation constrains the predictive capability of parametric models and typically restricts their performance to R
2 values of approximately 0.70. To overcome this shortcoming, recent years have witnessed a growing interest in artificial neural networks (ANNs) [
8,
9,
10,
11], support vector machines (SVMs) [
12,
13,
14,
15,
16,
17], fuzzy logic approaches [
18,
19,
20,
21], and hybrid soft computing techniques [
22,
23,
24,
25], owing to their flexibility in modeling nonlinear relationships and managing uncertainty, thereby enabling higher prediction accuracy in TBM performance estimation.
Recent studies have demonstrated that the performance of various machine learning and deep learning approaches strongly depends on site-specific conditions and dataset characteristics. For instance, Mahdevari et al. [
26] employed a support vector regression (SVR) model to predict TBM torque and total thrust with high accuracy, highlighting the critical role of parameter optimization (C, ε, γ). Ayawah et al. [
27] compared different machine learning models and emphasized that no single model consistently performs well across all field conditions, as prediction accuracy depends on both the number of input variables and site characteristics. Xu et al. [
28] reported that ensemble methods outperform statistical and deep learning approaches in terms of prediction accuracy when applied to limited datasets.
Li et al. [
29] utilized a dual-input 2D convolutional neural network (2D-CNN) to predict TBM torque, total thrust, and rock conditions, demonstrating that a “line model” emphasizing steady-state phase data yielded the best performance. Mahmoodzadeh et al. [
30] applied a long short-term memory (LSTM) model to a dataset comprising 1125 records and achieved the highest correlation coefficient and lowest mean squared error among the evaluated algorithms. Shahrour and Zhang [
31] recommended reducing input dimensionality, completing missing data, using a limited number of hidden layers, adopting recurrent neural networks for time-series data, and employing hybrid optimization techniques to mitigate the risk of convergence to local minima in soft computing applications.
Recent research underscores the critical importance of model selection, parameter optimization, and data preprocessing strategies in relation to dataset size and site conditions. While deep learning and ensemble methods generally provide superior performance—particularly in capturing complex relationships and handling limited datasets—they also suffer from several limitations, including data scarcity, site-specific constraints, high computational cost, and challenges in result interpretability.
Nevertheless, these methods are capable of predicting penetration rates with high accuracy under multivariate and complex geotechnical conditions, effectively capturing nonlinear relationships and intricate interactions among parameters. Despite achieving high predictive performance (R2 ≈ 0.90–0.95), the interpretability of model outputs and the ability to quantify the relative influence of input variables remain limited, posing a significant drawback for practical engineering applications.
To address this limitation, the present study develops a Jacobian-based framework to quantify the influence of model inputs on outputs by integrating parametric and machine learning approaches, and applies this framework to the prediction of TBM penetration rate (ROP). Within this context, six parametric models are considered: three conventional models (linear, exponential, and nonlinear) and three hybrid formulations. The parameters of all models are optimized using the Differential Evolution (DE) algorithm [
32,
33], and the resulting coefficients are employed to analyze the effects of input variables on the output through the Jacobian matrix.
The analyses are conducted under two distinct scenarios. In Scenario I, the model inputs are used in their original form, whereas in Scenario II, the inputs are standardized using the z-score method. For both scenarios, the relative importance of input variables is quantified and the results are presented comparatively. In addition, variable importance measures are computed for Random Forest (RF) [
34,
35], Bagged Trees (BT) [
36,
37], Support Vector Machine (SVM) [
38,
39], and Generalized Additive Models (GAM) [
40,
41], and the outcomes are compared with those obtained from parametric models.
For reliable modeling of TBM penetration rate, high predictive accuracy alone is insufficient; transparency and interpretability of model decision mechanisms are equally critical. Accordingly, Partial Dependence Plots (PDP) and Accumulated Local Effects (ALE) analyses are employed to investigate the influence of input variables on ROP within the selected machine learning models. As black-box models do not permit analytical differentiation, numerical differentiation techniques are applied. These analyses provide both global and local insights into variable effects, enhance model interpretability, and enable a reliable and comprehensive assessment of factor influence in TBM penetration rate prediction.
The main contributions of this study are summarized as follows:
- (1)
Development of a Jacobian-based framework to quantify the influence of input variables in both parametric and machine learning models for TBM performance prediction.
- (2)
Comparative analysis of variable importance across multiple modeling approaches under standardized and non-standardized input scenarios.
- (3)
Enhancement of model interpretability through PDP and ALE analyses, enabling more transparent and reliable assessment of factor influence in TBM penetration rate prediction.
The proposed framework is expected to improve the transparency of TBM performance prediction and support more informed decision-making in tunneling projects.
3. Model Application and Results
3.1. Implementation of Parametric Models (M1–M6)
In this section, the implementation process of the parametric models M1–M6 developed for predicting the TBM penetration rate (ROP) is presented. The study jointly evaluates both classical regression approaches (linear, exponential, and multiplicative nonlinear models) and hybrid nonlinear models supported by heuristic optimization techniques.
The parameters of the models defined in Equations (1)–(3) were estimated using conventional estimation methods. Accordingly, the Ordinary Least Squares (OLS) method was employed for the linear model [
36,
37], while parameter estimation for the exponential and multiplicative nonlinear models was performed using the Nonlinear Least Squares (NLS) approach [
38,
39,
40,
41].
In contrast, for the hybrid models expressed in Equations (4)–(6), parameter optimization was carried out using the Differential Evolution (DE) algorithm, which is capable of better adapting to nonlinear and complex data relationships [
26,
27].
The optimal model parameters obtained are presented in
Table 3, while the performance evaluations of the models based on various error metrics are comparatively summarized in
Table 4.
An examination of the results presented in
Table 4 indicates a consistent increase in the coefficient of determination (R
2) from M1 to M6, accompanied by a pronounced decrease in RMSE and MAE values. This trend can be attributed to the evolution of the model structures from linear and relatively simple formulations toward more complex nonlinear forms, allowing the relationships among variables to be represented more effectively.
While the independent variables in the initial models were expressed through linear or limited logarithmic and exponential transformations, the inclusion of exponential, logarithmic, and interaction terms in the subsequent models led to a substantial improvement in prediction accuracy. The M6 model, which was implemented for the first time in this study and explicitly incorporates interaction effects among variables, achieved the highest goodness-of-fit values and the lowest error metrics in predicting TBM performance.
It was observed that the Brazilian tensile strength (BTS) variable, which has often been neglected in previous studies, did not provide a significant contribution in Equations (1)–(3). In contrast, in the interaction-based Equations (4)–(6), BTS markedly improved model performance. To demonstrate the effectiveness of the Differential Evolution (DE) algorithm used for optimizing the parameters of the M1–M6 models, the parameters of the best-performing M6 equation were also estimated using commonly applied heuristic algorithms, including Harmony Search (HS), Particle Swarm Optimization (PSO), Symbiotic Organisms Search (SOS), and Grey Wolf Optimization (GWO). The comparative results are summarized in
Table 5.
As shown in
Table 5, the DE and SOS algorithms yielded the best performance across all error evaluation criteria, whereas the other algorithms became trapped in local optima. The convergence behavior of the algorithms for the M6 model is illustrated in
Figure 3a. As can be seen from
Figure 3a, no significant improvement in the results was observed after 100 iterations for any of the algorithms. However, as also evident from
Table 5, DE and SOS successfully converged to the global optimum, while the remaining algorithms stagnated at local optima, showing no noticeable improvement from the 100th to the 1000th iteration due to population similarity. Similar performance trends for the DE and SOS algorithms were also observed for the other models besides M6.
Figure 3 illustrates the convergence behavior of the M6 model, along with the variation in the model predictions and measured values along the tunnel length, and the corresponding scatter plot.
As shown in
Table 5 and
Figure 3, the M6 model stands out as the most successful among the investigated parametric models, achieving the highest coefficient of determination (R
2 = 0.69) and the lowest RMSE value (0.20). Moreover, the fact that its test performance exceeds that of the parametric models commonly reported as benchmarks in the literature [
4,
5,
6] demonstrates that the proposed model offers a robust and reliable alternative for predicting TBM penetration rate. In engineering applications, however, high predictive accuracy alone is not sufficient; understanding the relative contributions of the input variables to this performance is equally important. Accordingly, this study not only focuses on improving model accuracy but also quantifies the relative influence of the model inputs. This dual emphasis constitutes one of the most significant contributions of the present study to TBM performance modeling and is discussed in detail in the following section.
The proposed M6 model provides a parametric framework for estimating TBM penetration rate by explicitly incorporating key model variables, including BI, UCS, α, DPW, and BTS. The primary reason for its superior performance compared to other parametric approaches lies in the inclusion of the interaction term between BI and DPW. This interaction effectively captures the nonlinear mechanical behavior observed during penetration by jointly representing the cutter–rock contact mechanism, energy transfer, and stress redistribution processes. Although the calibrated model coefficients are site-specific, the inclusion of fundamental mechanical relationships—such as the BI × DPW interaction—allows the model structure to be adapted to different geological conditions, TBM types, and operational settings through recalibration or re-optimization using new field data. Consequently, the M6 model can be confidently employed as a field-calibrated predictive tool while also offering sufficient flexibility to serve as a transferable parametric modeling framework. This demonstrates that the model’s superior performance is not only physically grounded but also potentially generalizable.
3.2. Feature Importance Analysis for Parametric Models
As shown in
Table 1, the statistical characteristics and orders of magnitude of the variables constituting the model inputs differ substantially. For instance, uniaxial compressive strength (UCS) varies on the order of approximately 150 MPa, whereas the average spacing between discontinuities (DPW) is on the order of about 1 m. Such scale disparities may lead to misleading interpretations of the relative influence of variables, particularly in derivative-based sensitivity analyses.
For this reason, a Jacobian-based elasticity analysis was conducted under two different scenarios:
In both scenarios, the sensitivity of the model output with respect to the input variables was quantified using an elasticity formulation, and the resulting relative weights are presented comparatively in
Figure 4. This approach enables a scale-independent assessment and facilitates a more physically meaningful interpretation of the effects of the input variables.
An examination of the correlation matrix presented in
Figure 2 indicates that the strongest positive relationship with the penetration rate (ROP) is observed for the rock brittleness index (BI) (r = 0.5805), followed by uniaxial compressive strength (UCS) (r = 0.2595). The sensitivity results shown in
Figure 4 are consistent with these findings, demonstrating that BI and UCS exert the most dominant influence on ROP across all models, regardless of whether original or standardized data are used. This consistency confirms that the developed parametric models capture the variable–output relationships in a physically coherent and meaningful manner.
On the other hand, the relative importance of the discontinuity orientation (α) varies depending on both the data scaling and the model structure. While α exhibits a comparable weight across all models in the original-scale analyses, its influence becomes largely confined to the M1–M3 models when standardized data are used, diminishing markedly in the remaining models. As also indicated in
Table 2, this behavior stems from the wide range of variation exhibited by the α variable (min: 2°, max: 89°, mean: 44.715°, standard deviation: 23.279°). Such a broad distribution artificially inflates the magnitude of derivatives computed on the original scale, whereas standardization effectively mitigates this effect.
A pronounced negative relationship is observed between DPW and ROP (r = –0.4654), indicating that an increase in discontinuity spacing leads to a reduction in penetration rate and that DPW adversely affects TBM performance. PDP- and ALE-based analyses reveal that DPW exerts a moderate yet consistent influence, whereas the Jacobian (elasticity-based) analysis assigns relatively higher importance to DPW in certain models. This finding suggests that even small variations in DPW can induce meaningful local effects on ROP.
The Brazilian tensile strength (BTS) variable, by contrast, exhibits relatively low importance across all sensitivity analysis methods and models. This result indicates that the influence of BTS on ROP is secondary compared to other mechanical and structural parameters, in agreement with the outcomes of the correlation analysis.
The comparative results presented in
Table 6 clearly demonstrate that PDP and ALE methods capture the global average effects of the input variables, whereas the Jacobian (elasticity-based) approach reflects sensitivity based on local derivative information. Nevertheless, the strong consistency among the three methods in terms of the ranking of key variables confirms that the developed parametric models are reliable from both statistical and engineering perspectives.
3.3. Noise-Based Robustness and Variable Sensitivity Analysis for Parametric Models
To evaluate the behavior of the parametric models (M1–M6) developed in this study under measurement uncertainties and to assess the stability of variable effects, a noise-based robustness and sensitivity analysis was conducted. This analysis does not rely on re-estimation of model coefficients; rather, it aims to examine the responses of the final calibrated parametric models under perturbed input data. Accordingly, the applied methodology constitutes an input uncertainty propagation analysis, focusing on how measurement uncertainties propagate through the model to affect the outputs, rather than a re-fitting or re-training procedure.
Within the scope of the analysis, the coefficients of each parametric model were optimized once and subsequently held constant throughout the evaluation process. This ensures that the assessment exclusively reflects the impact of measurement errors in the input variables on the model outputs, while eliminating any effects arising from changes in model structure or parameter values.
To represent measurement uncertainties, multiplicative Gaussian noise was added to the model inputs. The noise injection process is defined as follows:
Here,
X denotes the base data matrix,
⊙ represents element-wise multiplication, and
ε∼N(0, σn2) denotes a random error term with zero mean and variance corresponding to the selected noise level. Accordingly, the perturbed data for each observation and variable are obtained as follows:
In this study, noise levels were set to 0.5%, 1%, and 5%. The primary reason for adopting a multiplicative noise approach is that measurement errors in geomechanical and rock mechanics parameters are predominantly scale-dependent (proportional) in nature. In particular, measurement uncertainties associated with parameters such as UCS, BI, and BTS typically manifest as a certain percentage of the measured value; this characteristic limits the physical consistency of additive error models. The multiplicative approach preserves the engineering meaning of the model inputs by preventing the occurrence of negative or physically meaningless values.
Following the noise injection process, the models were not retrained. All parametric models retained the optimized coefficients obtained from the noise-free dataset. This design choice enabled the analysis to isolate the effect of measurement uncertainty in the input variables on model outputs, without confounding influences arising from changes in model parameter stability.
To evaluate the sensitivity of model outputs to input variables, a Jacobian-based elasticity analysis was applied for each noise-perturbed scenario. For each observation, the elasticity value was computed as follows:
Here, denotes the model prediction obtained using the noise-perturbed inputs. The absolute means of the computed elasticity values were then calculated and normalized by dividing them by the total elasticity. This normalization procedure ensured that variables with different scales became directly comparable.
The results indicate that, as the noise level increases, the model prediction performance (e.g., RMSE) decreases as expected; however, the relative importance ranking of the variables is largely preserved. This finding demonstrates that, although parametric models experience a reduction in predictive precision under measurement uncertainty, they are still able to maintain the fundamental physical relationships among variables and the structural consistency of the model.
In this context, the applied noise-based analysis demonstrates that the proposed parametric models exhibit a robust structure not only in terms of statistical accuracy, but also with respect to interpretability, physical consistency, and engineering reliability.
Figure 5 illustrates how the relative sensitivities of input variables in the developed parametric models (M1–M6) vary under different noise levels. The panels represent cases with multiplicative Gaussian noise levels of (a) 0.5%, (b) 1%, and (c) 5%, respectively. All sensitivity values were computed based on the Jacobian-based elasticity analysis and subsequently normalized.
An examination of
Figure 5 reveals that, as the noise level increases, the fluctuations observed in model outputs and the associated prediction uncertainty also increase. Nevertheless, the relative importance ranking of the variables is largely preserved across all noise levels. In particular, the blockiness index (BI) and uniaxial compressive strength (UCS) consistently emerge as the dominant variables for all models and noise scenarios.
While the distribution of variable sensitivities exhibits a highly stable pattern under low (0.5%) and moderate (1%) noise levels, limited changes in relative weights are observed in some models at the 5% noise level. These variations are not sufficient to alter the underlying physical trends; notably, the overall hierarchy among the BI–UCS–DPW variables remains intact. This behavior indicates that the models are capable of maintaining structural consistency even under measurement uncertainty.
The relative contributions of the α and BTS variables display minor fluctuations as noise increases and do not become dominant at any noise level. This outcome suggests that the influence of these parameters on ROP remains secondary, and their sensitivity to measurement noise is relatively limited.
To facilitate a clearer interpretation of the trends presented in
Figure 5, the corresponding numerical values are provided in
Table 7. An examination of
Table 7 indicates that increases in the noise level lead to only limited fluctuations in elasticity values, while the relative importance ranking of the variables is largely preserved across all models. BI and UCS consistently emerge as the most dominant variables in all scenarios, whereas ALPHA and DPW exhibit moderate influence, and BTS shows a comparatively limited effect.
Although an increase in noise level (from 0.5% to 5%) results in a slight reduction in predictive sensitivity, it does not induce any structural disruption in model behavior. This observation demonstrates that the models are capable of preserving the fundamental physical relationships among variables as well as the relative hierarchy of their effects. Accordingly, the numerical results presented in
Table 7 provide strong evidence that the parametric models yield robust and interpretable outcomes not only in terms of accuracy but also under measurement uncertainty.
In summary, when the visual trends observed in the figures are jointly interpreted with the quantitative evidence provided in
Table 7, it becomes more evident which parameters exert strong influence and which require more cautious interpretation within the proposed modeling framework.
3.4. Feature Importance Analysis for Machine Learning Models
Feature importance analysis is a fundamental explainability approach that reveals which input variables a machine learning model relies on, and to what extent, when predicting the target variable. This analysis is critically important not only for assessing predictive performance but also for understanding through which variables this performance is achieved. In particular, for tree-based methods (such as Random Forest and Gradient Boosting), feature importance stands out as a powerful tool for interpreting the model’s decision-making mechanism [
36,
37,
38].
The primary objective of feature importance analysis is to determine the relative influence of variables in the dataset and thereby enhance the transparency of the model’s internal structure. In this way, the inputs that contribute most significantly to prediction accuracy can be clearly identified, substantially reducing the “black-box” nature of the model [
44].
In addition, feature importance analysis offers significant advantages in terms of dimensionality reduction and model simplification. By removing variables with low contribution or high collinearity, more parsimonious and interpretable models can be obtained, leading to reduced training time and computational cost while improving generalization performance [
44]. Moreover, this approach enables the comparison of variable priorities across different model types applied to the same dataset, allowing a clearer assessment of each variable’s sensitivity within the modeling framework.
Within the scope of this study, the effects of input variables used in modeling the TBM penetration rate (ROP) were examined in detail using four different machine learning methods (Random Forest, Bagged Trees, SVM, and GAM). Model performances under both original and standardized data scenarios were comparatively evaluated on a method-by-method basis and are presented in
Table 8.
An examination of
Table 8 reveals that, under both scenarios, the GAM exhibits a clearly superior performance compared to the other machine learning models, achieving the highest correlation coefficient (R) and coefficient of determination (R
2), along with the lowest error metrics (RMSE and MAE). In particular, the GAM attains R
2 = 0.906 and RMSE = 0.110 in Scenario I, indicating its strong capability to accurately capture the nonlinear relationships governing TBM penetration rate.
From a broader perspective, the Random Forest, Bagged Trees, and SVM models display only marginal performance variations with respect to data scaling (original versus standardized data). This behavior suggests that these models are relatively robust to scaling choices. In contrast, the use of standardized data in the GAM (Scenario II) results in a noticeable degradation in performance, reflected by a decrease in R2 and increases in RMSE and MAE. Nevertheless, despite this reduction, the GAM continues to outperform the other methods in both scenarios. This outcome highlights the high representational capacity of GAM and suggests that its performance may be sensitive to the original scale and distribution of input variables.
Following the evaluation of model performance across scenarios and methods, a feature importance analysis was conducted to identify which input variables contribute most significantly to predictive accuracy. The results of this analysis are presented in
Table 9 and
Figure 6. This analysis not only quantifies the overall predictive success of the models but also explicitly reveals, as further illustrated in
Figure 7, the extent to which individual variables contribute to that success.
Table 9 presents the relative importance of input variables obtained for the Random Forest (RF), Bagged Trees (BT), SVM, and GAM methods under Scenario I (original variables) and Scenario II (standardized variables). The findings indicate that the models produce largely consistent results across both scenarios and exhibit similar learning dynamics in terms of variable prioritization.
As evidenced by
Table 9 and
Figure 6, the BI and ALFA variables attain the highest importance scores across all models, emerging as the dominant inputs in predicting TBM penetration rate. In Scenario II, particularly for the SVM and GAMs, the distribution of variable importance becomes more balanced, with a noticeable increase in the relative contributions of UCS and DPW. In contrast, the tree-based models (RF and BT) exhibit only limited inter-scenario variation, with variable rankings largely preserved. This behavior indicates that tree-based approaches are less sensitive to variable scaling, reinforcing their robustness with respect to input data representation.
When
Figure 4,
Figure 6 and
Figure 7 are jointly examined, the observed differences between machine learning (ML) models and parametric models clearly indicate that the way variable contributions are reflected in the model outputs is strongly dependent on the model structure. In particular, although the ALPHA parameter exhibits a limited influence on the results of parametric models, it assumes a dominant role across all machine learning approaches. This contrast highlights a pronounced divergence between the two modeling paradigms in terms of their learning mechanisms and their capacity to capture nonlinear relationships.
Figure 7 demonstrates that ML methods exhibit a notable similarity in their data-driven learning strategies. Among these, GAM distinctly stands out due to its ability to capture local variations, achieving superior performance across all evaluation metrics. Furthermore,
Figure 8 clearly shows that while GAM and SVM display a more fluctuating learning process, Random Forest (RF) and Bagged Trees (BT) maintain a relatively stable performance throughout the training stages.
Similarly, the BTS variable—often treated as secondary in previous parametric studies and some ML approaches—provides a performance-enhancing contribution in both nonlinear parametric models and ML methods. This finding indicates that BTS plays a complementary yet meaningful role in explaining TBM performance.
Overall, the findings demonstrate that variable importance analyses yield consistent and reliable results across different scaling scenarios and model classes. Moreover, ML approaches exhibit a greater potential to represent the complex interactions influencing TBM penetration rate more effectively and transparently compared to parametric models.
3.5. Variable Contributions in Parametric and Machine Learning Approaches
To investigate variable contributions and robustness to noise across different model classes in predicting TBM penetration rate, a comprehensive analysis was conducted using both parametric and machine learning (ML) methods.
Table 9 presents the variable importance results for Random Forest (RF), Bagged Trees (BT), SVM, and Generalized Additive Model (GAM) under Scenario I (original data) and Scenario II (Z-score standardized data).
The results indicate that ML models produce largely consistent outcomes across both scenarios, exhibiting similar learning dynamics. Across all models, BI and ALPHA emerge as dominant inputs in predicting TBM penetration rate. However, when standardized data are used, particularly in SVM and GAM, variable importance distributions become more balanced, with UCS and DPW showing notable increases in their relative contributions. In contrast, tree-based models (RF and BT) preserve their variable rankings across scenarios, demonstrating limited sensitivity to variable scaling.
The comparison between parametric and ML models reveals that the way variables are incorporated into the model is closely linked to the underlying model structure. While ALPHA exhibits a relatively limited effect in parametric models, it assumes a dominant role in ML models. This difference underscores the ability of ML approaches to learn nonlinear interactions and complex relationships more effectively than parametric formulations. Similarly, while BTS plays a secondary role in parametric models, it significantly enhances model performance in ML methods—particularly in GAM and SVM—thereby assuming a complementary role in explaining TBM performance.
Noise robustness analyses (
Table 7) further provide valuable insights. The Jacobian-based normalized elasticity values obtained under different multiplicative Gaussian noise levels (0.5%, 1%, and 5%) indicate that the models largely preserve their fundamental variable priorities. BI and ALPHA remain dominant across all models and noise levels, while other variables exhibit only limited fluctuations. This finding confirms that the models retain a considerable degree of robustness to measurement uncertainty when predicting TBM penetration rate.
In summary, variable importance analyses deliver consistent and reliable insights across scaling scenarios and model classes. However, structural differences among models play a critical role in interpreting variable contributions. Parametric models are particularly effective in representing fundamental relationships within a predictable mathematical framework, but their capacity to capture nonlinear and complex interactions is limited. In contrast, ML models—especially GAM and SVM—are more capable of capturing such complexity, modeling interactions among variables that influence TBM penetration rate more effectively. This demonstrates that the two approaches are complementary, with ML methods offering superior potential when complex field data and noise robustness are of concern.
4. Discussion
The variable importance analyses conducted using both parametric and machine learning (ML) models provide a multidimensional perspective on the factors influencing TBM penetration rate (ROP). Because input variables in parametric models exhibit markedly different scales (e.g., UCS ≈ 150, DPW ≈ 1), direct comparison of variable effects is challenging. Therefore, Jacobian-based elasticity analysis was performed under two scenarios: original-scale variables (Scenario I) and standardized variables (Scenario II). The resulting relative variable weights are comparatively presented in
Figure 3, allowing assessment of scale effects on the results.
All coefficients of the six parametric models (M1–M6) were optimized using the Differential Evolution (DE) algorithm. Consequently, variable importance evaluations were based not on fixed or literature-derived coefficients, but on model-specific parameter sets optimized for predictive performance. This approach enhanced the reliability of parametric analyses by ensuring consistency between variable interpretation and model accuracy.
Parametric analyses reveal that the strongest positive relationship with ROP is associated with the Blockiness Index (BI, r = 0.5805), followed by Uniaxial Compressive Strength (UCS, r = 0.2595). The joint angle (α) exhibits a relatively limited effect (r = 0.2196), while DPW shows a pronounced negative relationship with ROP (r = −0.4654). The dominant influence of BI and UCS is consistently preserved across all parametric models and both scenarios, in agreement with correlation matrix findings. Scenario-related differences reflect redistribution effects caused by scaling rather than changes in physical influence.
Among the parametric formulations, the proposed M6 model provides a more balanced and physically meaningful distribution of variable contributions. Elasticity analysis based on DE-optimized M6 coefficients not only confirms the dominance of BI and UCS, but also highlights indirect and interaction-based contributions from DPW and BTS. This indicates that the interaction and logarithmic terms embedded in M6 more realistically represent TBM–rock interactions. Accordingly, M6 achieves higher predictive accuracy than existing parametric correlations reported in the literature.
Variable importance analyses for ML models (RF, BT, SVM, GAM) also yield largely consistent results across both scenarios (
Figure 4,
Table 6). The key distinction between parametric and ML models lies in how variable contributions are embedded within the model structure. ALPHA, which has limited influence in parametric models, attains high importance in ML models—demonstrating ML methods’ superior ability to capture nonlinear and data-driven interactions. Similarly, BTS, while secondary in parametric models, contributes significantly in ML approaches—particularly in GAM, reaching importance levels of 14–15%.
Noise analyses presented in
Table 7 (0.5%→5%) further demonstrate model robustness. Although increasing noise introduces minor fluctuations in predictions, relative variable rankings remain preserved. BI and UCS consistently remain dominant, ALPHA and DPW maintain moderate influence, and BTS continues to play a secondary role. This confirms that the models can reliably represent both linear and nonlinear relationships even under measurement uncertainty.
Performance comparisons confirm that Model M6 outperforms all other models in terms of both error metrics (RMSE, MAE) and explanatory power. Compared to Model 1, M6 achieves a 9.07% reduction in RMSE and a 10.48% increase in R2. Similar improvements across other models indicate that, despite its increased complexity, M6 does not exhibit overfitting and delivers stable, balanced predictions—making it suitable as a reference model for subsequent analyses.
Overall, the combined use of parametric and ML models enables a comprehensive assessment of variable effects on TBM penetration rate, integrating both physical interpretability and nonlinear interaction modeling. While parametric models—particularly the DE-optimized M6—provide a transparent and physically consistent representation of processes, ML models successfully capture complex patterns embedded in the data. Together, these approaches form a complementary and robust modeling framework for TBM performance prediction.
5. Conclusions
In this study, six parametric models (M1–M6) were developed to predict the TBM penetration rate (ROP), with all coefficients optimized using the Differential Evolution (DE) algorithm. Model performance was evaluated using optimized coefficients, and variable importance was systematically examined through Jacobian-based elasticity analyses under both original (Scenario I) and standardized (Scenario II) data. This integrated framework enabled a consistent assessment of predictive accuracy, robustness, and scale sensitivity.
The comparative analyses demonstrate that the proposed M6 model outperforms widely used parametric formulations reported in the literature. Owing to the inclusion of interaction and logarithmic terms, M6 provides a more realistic representation of TBM–rock interactions and achieves superior predictive performance, yielding the lowest error metrics (RMSE, MAE) and the highest explanatory power (R2) among the evaluated parametric models.
Elasticity analyses consistently identify brittleness index (BI) and uniaxial compressive strength (UCS) as the dominant factors governing ROP, while mean spacing of weakness planes (DPW) exhibits a clear negative influence. The similarity of elasticity trends across both scenarios indicates that data scaling does not alter the underlying physical effects but rather redistributes relative contributions. The interaction terms incorporated in M6 further enhance the model’s ability to capture complex inter-variable relationships that cannot be adequately represented by linear formulations.
Machine learning (ML) models, including RF, BT, SVM, and GAM, yield results that are broadly consistent with the parametric findings. In particular, the increased prominence of the joint orientation angle (α) in ML models highlights their capability to capture nonlinear effects and complex dependencies. Moreover, although Brazilian tensile strength (BTS) plays a secondary role in classical parametric models, its inclusion contributes measurably to prediction accuracy, especially in GAM and nonlinear parametric formulations.
Noise analyses conducted under varying noise levels (0.5–5%) confirm the robustness of the proposed framework, as relative variable importance rankings remain stable. BI and UCS consistently retain their dominant influence, while α and DPW exhibit moderate effects and BTS maintains a limited yet meaningful contribution, supporting the physical interpretability and engineering relevance of the results.
Overall, the results demonstrate that the proposed M6 parametric model, optimized using Differential Evolution and supported by Jacobian-based elasticity analysis, achieves the most balanced performance among all evaluated approaches. While machine learning models such as GAM and SVM effectively capture nonlinear interactions, the M6 model offers superior interpretability, physical consistency, and competitive predictive accuracy. The hybrid evaluation framework adopted in this study therefore provides a robust and practically applicable methodology for reliable TBM penetration rate prediction under complex geological conditions.
Future studies may extend the current framework by comparing Jacobian-based elasticity results with model-agnostic explainability techniques such as SHAP or permutation importance. Such a comparison would strengthen the interpretability of the proposed approach by highlighting both local and global variable contributions.