Next Article in Journal
A Research Paper on the Influence of Blast Weakening on the Vibrations of Ground Buildings in a Shallow-Buried Extra-Thick Coal Seam
Previous Article in Journal
1H-qNMR as a Tool for the Quantitative and Qualitative Evaluation of Abietane-Type Diterpenes in Lamiaceae Species Cultivated in Greece
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Application of Extra-Trees Regression and Tree-Structured Parzen Estimators Optimization Algorithm to Predict Blast-Induced Mean Fragmentation Size in Open-Pit Mines

School of Resources and Safety Engineering, Central South University, Changsha 410083, China
*
Authors to whom correspondence should be addressed.
Appl. Sci. 2025, 15(15), 8363; https://doi.org/10.3390/app15158363
Submission received: 8 June 2025 / Revised: 19 July 2025 / Accepted: 24 July 2025 / Published: 28 July 2025
(This article belongs to the Section Computing and Artificial Intelligence)

Abstract

Blasting is an effective technique for fragmenting rock in open-pit mining operations. Blasting operations produce either boulders or fine fragments, both of which increase costs and pose environmental risks. As a result, predicting the mean fragmentation size (MFS) distribution of rock is critical for assessing blasting operations’ quality and mitigating risks. Due to the limitations of empirical and statistical models, several researchers are turning to artificial intelligence (AI)-based techniques to predict the MFS distribution of rock. Thus, this study uses three AI tree-based algorithms—extra trees (ET), gradient boosting (GB), and random forest (RF)—to predict the MFS distribution of rock. The prediction accuracy of the models is optimized utilizing the tree-structured Parzen estimators (TPEs) algorithm, which results in three models: TPE-ET, TPE-GB, and TPE-RF. The dataset used in this study was collected from the published literature and through the data augmentation of a large-scale dataset of 3740 blast samples. Among the evaluated models, the TPE-ET model exhibits the best performance with a coefficient of determination (R2), root mean squared error (RMSE), mean absolute error (MAE), and max error of 0.93, 0.04, 0.03, and 0.25 during the testing phase. Moreover, the block size (XB, m) and modulus of elasticity (E, GPa) parameters are identified as the most influential parameters for predicting the MFS distribution of rock. Lastly, an interactive web application has been developed to assist engineers with the timely prediction of MFS. The predictive model developed in this study is a reliable intelligent model because it combines high accuracy with a strong, explainable AI tool for predicting MFS.

1. Introduction

Rock fragmentation plays a pivotal role in engineering disciplines such as mining, civil engineering, and construction [1,2,3,4]. In mining engineering, the rock fragmentation process entails the disintegration of rock mass or ore into smaller fragments using explosives, drilling apparatuses, and thermal stress. Studies have indicated that only approximately 30% of the explosive energy is effectively used for fragmentation [5,6]. The remaining energy is disseminated into other forms, which leads to adverse effects including back break, airblast, fly rock, and ground vibration [7,8,9,10,11,12,13]. The fragmentation size is the most critical factor during the blasting process, as it directly influences the efficiency of ore extraction, downstream processes, equipment management, and material handling [14,15]. Achieving an ideal fragmentation size not only enhances mines’ productivity but also improves their operational safety. Therefore, a comprehensive understanding of the fragmentation process is essential for making accurate predictions and optimizing outcomes.
Predicting the optimum fragmentation size is a challenging task due to the inherent complexity of blasting techniques [16,17]. To address this, researchers have proposed various approaches to optimize parameters and simplify the problem. For example, Kuznetsov developed a model to predict the mean fragmentation size (MFS) based on the properties of the rock mass and explosives used [18]. Cunningham subsequently refined this model, known as the Kuz–Ram model, by incorporating specific energy and additional rock mass characteristics [19,20,21]. Although these empirical models have limitations, they laid the groundwork for subsequent advancements in the field. Recognizing the need for more accurate predictions, researchers developed models that consider the full range of blasted fragment sizes, such as the distribution-free, Swebrec, and Rosin-Rammler functions [14,22,23,24]. Moreover, improvements to the Kuz–Ram model, such as the modified Kuz–Ram, Bergmann, and Riggle models, were proposed to enhance its predictive accuracy [25,26]. To further improve its practicality, researchers also developed equations for the BRW model to facilitate its real-world application.
Recently, researchers have progressively embraced artificial intelligence (AI)-based methods because of the limitations of the empirical models, especially their reliance on labor and costly field blast experiments. AI-based methods have achieved significant success in the geotechnical domain. For instance, AI-based methods have improved the precision of rock fragmentation predictions, pattern recognition, and data analysis [27,28,29,30,31]. Commonly used AI-based techniques for predicting blast-induced rock fragmentation include artificial neural networks (ANNs), fuzzy logic, genetic algorithms, and ensemble methods. Table 1 provides an overview of the AI-based techniques applied in the prediction of MFS. However, these techniques still need improvements, such as a reduction in their training time and improvement in their prediction accuracy, regardless of the quantity of the dataset. Among these techniques, tree-based algorithms have shown superior performance compared to ANNs, etc., even without extensive fine-tuning [32,33].
The existing literature reveals that most prior studies have primarily relied on small-scale blasting datasets, which limit models’ performance and increase the risk of overfitting. In contrast, AI-based models exhibit significantly improved performance when trained on large-scale datasets, as these enable the models to capture more complex patterns within the data. Moreover, large-scale datasets also improve the models’ generalization and allow them to perform better on new data.
Therefore, this study has the following objectives: (1) to employ a large-scale database for better model training; (2) to compare the predictive accuracy of the three tree-based models—extra trees (ET), gradient boosting (GB), and random forest (RF)—in predicting the MFS distribution of rock; and (3) to develop an interactive web application for easy and timely predictions. The remainder of this paper is structured as follows: Section 2 provides an in-depth description of the data source, data collection and data supplementation. Section 3 delves into the methodologies employed, including details on AI tree-based models, the Bayesian optimization algorithm, and Shapley additive explanations (SHAP), which were used for model interpretation. Section 4 presents the proposed model development and the statistical evaluation indices that were employed. Section 5 discusses the results obtained from comparative analysis of the models, hyperparameter optimization, and the SHAP model’s interpretability and provides a link to the rock fragmentation web application. Lastly, Section 6 concludes by summarizing the key findings and this study’s contributions, and outlines this study’s limitations and future research directions.

2. Data Source

2.1. Data Collection and Supplement

This study collected 187 blasting records from the published literature on in situ operations, as shown in Table 2. The MFS was represented by X50. To ensure consistent input parameters, the data collected by Sharma et al. [53] required supplementation due to missing values for two parameters: the elastic modulus (E) and in situ block size (XB). The elastic modulus was derived from the uniaxial compressive strength using Equation (1) [54], while the in situ block size was estimated by Bieniawski using the powder factor, spacing, burden, muck pile’s mean fragmentation size, and rock mass rating (RMR) with Equations (2) and (3) [36].
E = 0.3752 × UCS + 4.479
FI = 0.03 × ( RMR × PF × SB ) + 0.73
X B = FI × X 50
where the units of E, UCS, and Pf are GPa, MPa, and kg/m3, respectively, RMR is Bieniawski’s rock mass rating, and FI is the Fragmentation Index. The FI is estimated using Equation (2), and the XB is calculated using Equation (3). The RMR of the rock excavated by Sharma et al. [55] is taken as 50, which belongs to the ‘Fair’ class [56]. After supplementing the data, they are then converted to consistent input parameters with other datasets from Hudaverdi [57] and Zhu [58] as required for this study. As a result, a small-scale dataset of 187 groups of in situ blasting data with seven variables including the spacing/burden ratio (S/B), bench height/burden ratio (H/B), burden/hole diameter ratio (B/D), stemming/burden ratio (T/B), PF, XB, and E was combined with fragmentation size data to generate a complete database.

2.2. Data Augmentation

Data from various sources may present inconsistencies, noise, incompleteness, or inadequate quantity, which can lead to challenges in data analysis. This study’s small-scale database signifies that the fragmented data may be considered incomplete. The acquisition of blasting data is time-consuming and expensive, which has led to primarily limited-scale datasets. Inadequate data diminish the validity of insights that are obtained and decrease the accuracy of models in the training process. A substantial quantity of data is essential to accurate predictions and analyses, as well as for preventing overfitting and ensuring robust models. For tree-based algorithms, the prediction accuracy depends not only on hyperparameter optimization but also on the quantity of training data. Larger training datasets enable ML algorithms to better capture patterns and relationships, increasing their likelihood of making accurate predictions on unseen data. Moreover, acquiring blasting data is both time-consuming and costly [52]. Consequently, data augmentation offers a practical and recommended approach to generating large-scale datasets and thereby enhances the performance and generalizability of ML models.
Data augmentation methods vary depending on the type and dimensionality of the data. For instance, techniques such as rotating, cropping, and filtering are commonly used for image data, while jittering, slicing, and permutation are employed for curve data [59,60]. Among these, only filtering and jittering are noise-based methods. Noise-based methods are widely adopted for data augmentation because they are versatile and applicable across different data types [61]. This study implemented noise-based data augmentation by adding noise that was equivalent to 3% of the parameter values and that followed a normal distribution. This level of noise is designed to mimic uncontrollable human errors typically associated with the measurement and collection of in situ blasting data [52]. Additionally, the 3% noise does not present significant differences between the original and the augmented dataset. The dataset is organized as a 187 × 8 matrix that contains 187 blasting samples represented as rows, with seven input parameters and one output parameter represented as columns. To expand the dataset, each data point was used to generate 20 additional points through noise-based augmentation. The newly generated data points follow a normal distribution, with the mean set being equal to the original value and the standard deviation being adjusted so that 99.7% of the generated points fall within three standard deviations of the mean. This process was applied to every value in the matrix, which resulted in a significantly larger dataset of 3740 data groups, with each retaining the same seven input and one output parameter structure. A statistical summary of the augmented large-scale dataset is provided in Table 3.
Figure 1 presents a correlation heatmap of the features in the dataset. The correlation was computed using the mutual information (MI) regression technique, a measure of the shared information shared between two variables. MI regression quantifies how much knowing one variable reduces the uncertainty of another variable. The MI value is equal to zero if two variables are independent, and higher values indicate stronger dependency. MI is computed using Equation (4). Unlike the Pearson correlation, the use of the MI approach can capture non-linear dependencies without assuming any specific data distribution. Given the non-linear nature of the dataset, MI regression is suitable for computing the correlation between variables. As shown in Figure 1, four features exhibit a strong relationship with X50, i.e., BD, XB, HB, and E.
I X ; Y = p x , y log p x , y p x p y d x d y
where I ( X ; Y ) denotes the mutual information between X and Y. p ( x , y ) denotes the joint probability density function (pdf) of X and Y. p ( x ) , p ( y ) denotes marginal pdfs of X and Y.

3. Methodology

3.1. Random Forest Algorithm

RF is a supervised machine-learning algorithm that leverages multiple decision trees to deliver more accurate prediction. As an ensemble learning method, RF combines the predictions of multiple decision tree models and outputs the mean prediction for regression tasks [62,63,64]. The RF regressor operates an estimator that fits a collection of N randomized decision tree regressors independently and identically distributes random vectors [53]. Each decision tree contributes a unit vote for the most popular class or value at a given input x, which ensures diversity in the training process. Figure 2 shows the framework of the RF algorithm.
In RF training, the first step is to create random subsets of a given training dataset of independently and identically distributed 0 , 1 d × R -valued random variables (d ≥ 2) with a similar distribution as independent generic pairs (x, y) by randomly resampling a certain percentage of the total dataset t n = x 1 , y 1 , , x n , y n . This process is known as the bagging approach, and it uses multiple decision trees (DTs) without pruning [65,66]. The second step is randomly choosing a K input for lead node splitting while training each DT. The output prediction of the n-th decision tree is denoted as g ( x , t N n ) , and the average of the combined trees, taken to form a finite forest predictor, is defined in Equation (5):
y ^ = 1 n i = 1 n g x ,   t m n
where y ^ is the RF’s final prediction.

3.2. Extra Trees Algorithm

The extra trees (ET) algorithm proposed by Geurts et al. [67] is an example of an ensemble-based algorithm derived from the RF algorithm. Two key differences between ET and RF are as follows: (1) the ET builds trees using a random subset of the data without replacement, which helps minimize bias, and (2) the randomness in the ET comes from random splits of data rather than bootstrap aggregating [67]. In the ET method, the tree splitting process is regulated by two parameters: k is the number of nodes randomly chosen in the node, and nmin is the minimal size of the sample required to separate the node. These enhance model accuracy and help prevent overfitting in the ET algorithm [68,69].
Similar to the RF algorithm, the ET algorithm performs regression in two stages: bootstrapping and bagging. During the bootstrapping phase, a random sample of the training dataset is selected to build the decision trees. In the bagging phase, the nodes of the decision are divided using the random subsets of training data. The final decision is made by randomly selecting a subset and determining its corresponding value. Figure 3 illustrates the framework of the ET algorithm. Mathematically, the ET regression process is expressed by the equation provided by Breiman [70].
p x , β 1 , , β n = 1 m i = 1 m p x , β n
where p is the p-th prediction tree. β is a distributed vector.

3.3. Gradient Boosting Algorithm

In contrast to bagging, the boosting approach generates base models sequentially. This method improves prediction accuracy by producing models in succession, with each new model focusing on the training samples that are more difficult to predict. During the boosting process, the training data for each successive model emphasize the samples that were misclassified by previous base models, rather than those that were correctly predicted. Each new base model corrects the errors made by its predecessors [71,72].
The boosting approach resamples the training data to provide optimal information for each model. In each training step, the data distribution is adjusted based on the errors made by previous models. Emphasis is placed on samples that were inaccurately predicted, which gives them greater weight in subsequent models [73]. Boosting involves fitting additional models that minimize a loss function, which is averaged over the training data. This loss function quantifies the deviation between the predicted and true values. The process uses a forward stage-wise modeling approach, where new base models are added sequentially without the parameters or coefficients of the previously fitted models being altered [74]. The stage-wise steps for the gradient-boosting regression tree method are as follows:
Firstly, the constant value is initialized using Equation (7).
F 0 x = argmin α i = 1 m L ( y i , α )
where L ( y i , α ) is the loss function.
Secondly, the residuals can be calculated using Equation (8) as follows:
r i p = L y i , F x i F x i F x = F p 1 ( x ) f o r   i = 1 , , n
where rip is each sample’s residuals.
Thirdly, a terminal node is created, and a mean function that minimizes the loss function is found using the following equation:
α j p = argmin α x i R j p L y i , F m 1 x i + α   f o r   j = 1 , , J p
Finally, the model is updated using Equation (10):
F p x = F p 1 x + v j = 1 J p α j p 1 x R j p

3.4. Optimization Algorithm: Bayesian Optimization

The TPE is a Bayesian optimization approach proposed by Bergstra et al. [75]. TPEs are better than other optimization algorithms because (1) they handle high-dimensional hyperparameter search spaces; (2) they converge faster than most optimization algorithms such as meta-heuristic algorithms, grid search, and random search; (3) they are efficient for training computationally expensive machine learning algorithms; and (4) they are easy to implement with different libraries in python. To that end, three predictive models are constructed and cross-validated five times during training to assess the different models’ generalization capabilities. The TPE utilizes the Gaussian mixture model to find the best hyperparameter impacting the models [76,77]. The TPE runs on two main ideas: (1) using Parzen estimators to model the best hyperparameters for the model and (2) using a tree-like data structure known as the posterior-inference graph to optimize the algorithm runtime. Let X be a tree-structured search space and f : X be an objective function, then the minimization problem is defined as γ argmin γ X f γ . Suppose a set of observations is T = γ ( 1 ) , y ( 1 ) , , γ n , y n , then the TPE replicates p γ y for every parameter γ X i using two probability density functions as follows:
p γ y = l ( γ ) i f   y <   y g γ i f   y y
where p γ y is the conditional probability of the hyperparameter γ and the model loss y; l γ establishes the subset of the observed probability density values that are less than y * , and g ( γ ) establishes the subset of the remaining observed probability density values; the value y is chosen to be a quantile λ 0 , 1 of the observed y values that satisfy p ( y < y ) = λ .
To choose a particular candidate for evaluation, the TPE employs the expected improvement (EI) for y and the parameter γ as the acquisition function.
E I y ( γ ) = y max y y , 0 p y γ d y = y y y p y γ d y α λ + g γ l γ 1 λ 1
To gain new information with the maximum EI, g ( γ ) / l ( γ ) is minimized. To keep minimizing g ( γ ) / l ( γ ) , the process is repeated where γ is returned in the function and fitted again until the hyperparameter optimization is complete. It can be concluded that the probability of l ( γ ) is as high as possible, while the probability of g ( γ ) is as low as possible when the improvement is maximized [78].

3.5. Shapley Additive Explanations for Model Interpretation

The AI-based methods are often considered unreliable in many fields due to their opacity and insufficient explanation for the predictions rendered. The internal workings of these models are complex and difficult to comprehend using human logic, which has earned them the label of ‘black box’ models [79]. To address this issue and enhance transparency, explainable AI (XAI) tools are employed to interpret these models and provide insights into the complex relationships between input features and output predictions [80]. These XAI tools include SHAP and local interpretable model-agnostic explanations (LIME). In this study, SHAP is used to interpret the models because of the following reasons: (1) it provides a global understanding of the dataset, providing insights into the overall contribution of each feature [81]; (2) it has great visualizations that make it easier to understand the model’s behavior; and (3) it handles feature interactions, showing the impact of two or more features when combined [82].
SHAP is a cumulative feature attribution algorithm that assigns a relevance score to each input feature. It leverages concepts from game theory to calculate feature importance and satisfies key properties of feature attribution, including accuracy, missingness, and consistency [83]. Given a dataset D, the Shapley value estimation of an m-th feature with i combinations of features, target feature y, and predictive model f is computed using Equation (13).
ϕ i m = f ^ ( y + m ) f ^ ( y m )
where ϕ i m is the mean Shapley value of the m-th feature with the i-th feature. f ^ ( y + m ) is the prediction for target feature y with a random number of feature values, which also includes the m-th feature. f ^ = ( y m ) is the prediction without the m-th feature. Generally, Equation (14) is used to compute the Shapley value of the m-th feature.
ϕ m ( y ) = 1 n i = 1 n ϕ i m

4. Model Development and Evaluation Indices

This study employed three tree-based ensemble models to predict the MFS distribution of rock: RF, ET, and GB. The accuracy of these models depends on the optimal combination of hyperparameters. To identify this optimal combination, tree-structured Parzen estimators (TPEs) were used for hyperparameter tuning. Consequently, three hybrid models were developed: TPE-RF, TPE-ET, and TPE-GB, as illustrated in Figure 4.
The optimization process for all three models follows a similar and consistent procedure:
(1)
Data preprocessing: The dataset is divided into training and test sets with 75% (2805 samples) being allocated for training and 25% (935 samples) being reserved for testing. This 75/25 data split was chosen for several reasons: (1) it provides enough samples for the models to learn patterns, which prevents underfitting; (2) smaller test sets tend to be biased while larger test sets offer a more representative sample; and (3) this split ratio generalizes better than a 70/30 split or 80/20 split. Figure 5 show the data distribution between the training and test datasets. It can be observed that the training and test sets are normally distributed; therefore, there is no need for further pre-processing. Additionally, the training and test sets are distributed the same;
(2)
Optimization process: Many influencing parameters can influence the performance of the algorithms, but only a limited number of hyperparameters were selected to balance performance and computational cost. Five hyperparameters were optimized for the RF and ET algorithms, while six were used for the GB algorithm, as shown in Table 4. These hyperparameters are critical for improving the accuracy of the models. The performance of the Bayesian optimization algorithm is influenced by the number of trials or iterations used to search for the optimal hyperparameter combinations. More trials result in longer training times and higher costs, whereas fewer trials may lead to underfitting. To maintain consistency across models, the number of trials was set to 150 for all models;
(3)
Model evaluation: In this study, four indices, including the root mean squared error (RMSE), mean absolute error (MAE), coefficient of determination (R2), and max error were employed to evaluate the performance of the three models. These metrics are described using Equations (15)–(18):
RMSE = i = 1 n y ^ i y i 2 n
MAE = 1 n i = 1 n y ^ i y i
R 2 = 1 i = 1 n y i y ^ i 2 i = 1 n y i y ¯ i 2
Max   Error ( y , y ^ ) = max ( y i y ^ i )
where y i represents the actual value, y ¯ i represents the mean value of the predicted target, and y ^ i represents the predicted value.
Figure 5. Data distribution between training and test datasets: (A) data distribution of SB; (B) data distribution of HB; (C) data distribution of BD; (D) data distribution of TB; (E) data distribution of Pf (kg/m3); (F) data distribution of XB (m), and (G) data distribution of E (GPa).
Figure 5. Data distribution between training and test datasets: (A) data distribution of SB; (B) data distribution of HB; (C) data distribution of BD; (D) data distribution of TB; (E) data distribution of Pf (kg/m3); (F) data distribution of XB (m), and (G) data distribution of E (GPa).
Applsci 15 08363 g005
Table 4. Hyperparameters and their search space.
Table 4. Hyperparameters and their search space.
HyperparametersRegression Algorithms Hyperparameter Search SpacesData TypeDescription
TPE-ETTPE-RFTPE-GB
n estimators[10, 3000][10, 3000][10, 3000]IntegerNumber of trees in the forest
max depth[2, 40][2, 40][2, 40]IntegerMaximum depth of each tree
min samples split[2, 35][2, 35][2, 35]IntegerThe minimum number of samples required to split an internal code
criterion‘squared_error’, ‘absolute_error’, ‘friedman_mse’, ‘poisson’‘squared_error’, ‘absolute_error’, ‘friedman_mse’, ‘poisson’‘squared_error’, ‘friedman_mse’,CategoricalThese functions measure the quality of a split.
min impurity decrease[0.00001, 0.9][0.00001, 0.9][0.00001, 0.9]FloatIt splits the node if the split induces a decrease of the impurity greater than or equal to this value.
learning rate [0.00001, 0.9]FloatIt shrinks the contribution of each tree by the value of learning_rate.

5. Results and Discussion

5.1. Performance Comparison of Models for MFS Prediction

The three optimized models were compared to determine the best model for predicting the MFS distribution of rock. As outlined in Section 4, aside from the TPE-GB model, the number of trials and hyperparameters for each model was kept consistent to ensure fairness in the results. Figure 6 presents a bar graph that shows the importance values of different hyperparameters during the optimization process. Figure 6 illustrates that all models prioritized the ‘criterion’ hyperparameter, which is rational, as it assesses potential splits at each node to optimize information gain. In the TPE-RF model, the ‘criterion’ hyperparameter significantly influenced the optimization process, whereas the other hyperparameters had a minimal impact. Figure 7 displays the performance of the three models during the optimization process involving 150 trials. The results indicate that the performance differences among the models were not substantial. However, the TPE-ET model exhibits better performance than the other two models. Additionally, it was observed that some hyperparameter combinations during the optimization process resulted in negative values of the objective function, as shown in the subfigure in Figure 7.
According to Table 5, the TPE-GB model outperformed the other two models during the training phase, achieving R2, RMSE, MAE, and max error values of 0.97, 0.03, 0.02, and 0.14, respectively. However, as shown in Table 6, the TPE-GB model underperformed compared to the other models during the testing phase. In contrast, the TPE-ET model demonstrated superior performance during the testing phase, with R2, RMSE, MAE, and max error values of 0.93, 0.04, 0.03, and 0.25, respectively. However, the evaluation metrics across the models are closely matched, with minimal differences, which makes it challenging to definitively identify the best-performing model. The lower max error values suggest that TPE-GB and TPE-ET were the most reliable during the training and testing phases, respectively. To further analyze and compare the model performances, score rankings were visualized using stacked bar charts. As illustrated in Figure 8a,b, the TPE-GB model achieved the highest score during the training phase but ranked lower during the testing phase. Meanwhile, the TPE-ET and TPE-RF models exhibited comparable scores during the testing phase.
To further understand the models’ performance, regression plots are utilized to explore the relationship between the predicted X50 values and actual X50 values. Figure 9 and Figure 10 depict scatter plots of the three models, where the predicted X50 is on the y-axis and the actual X50 is on the x-axis. The red dashed line represents the best-fit line, indicating a perfect relationship between the predicted and the actual values. As shown in Figure 9a–c, all models demonstrate a strong relationship between the predicted and actual values during the training phase. Among them, the TPE-GB model exhibits a slightly better relationship with the best-fit line than the other models and achieves the highest R2 value. Similarly, Figure 10a–c depict the relationship between the predicted and the actual values during the testing phase. All models exhibit a strong correlation between the predicted and actual values, although several outliers significantly deviate from the best-fit line, as indicated by the green circles in the figures.

5.2. Model Interpretation

Figure 11 presents a bee swarm plot that illustrates the distribution of feature contributions alongside the SHAP interaction matrix. The matrix’s diagonal elements indicate the independent contribution of each feature to the model’s predictions, whereas the off-diagonal elements reflect the interaction effects between pairs of features. The SHAP value is derived by summing the independent contributions of individual features and their interaction contributions. The figure indicates that XB(m) significantly influences the model’s predictions, with a notable interaction between XB(m) and E(GPa) enhancing this predictive effect. Conversely, SB demonstrates minimal independent contribution to the model’s predictions, and the interaction between SB and BD exhibits a negligible effect on the model’s output.
Figure 12 integrates a local bar plot with a bee swarm plot to provide a comprehensive visualization of feature contributions to the model’s predictions. The bars in the local bar plot represent the SHAP values for each feature, illustrating the individual contributions of each. The top x-axis corresponds to the mean SHAP values from the bar plot, while the bottom x-axis represents the SHAP value contributions in the bee swarm plot. In the bee swarm plot, the color coding indicates feature value ranges, with blue signifying low values and red signifying high values. It can be observed that XB plays a significant role in the model’s prediction. High values of XB positively influence the predictions, whereas low values have a negative impact. In contrast, SB has minimal influence on the model, with high values of SB contributing slightly positively to the predictions.
Figure 13 employs a scatter plot with a LOWESS fit curve to provide an intuitive understanding of how the range of feature values influences the model’s predictions. The LOWESS curve (the red curved line) represents a weighted regression curve that smooths out non-linear trends in the data. Figure 13 shows the fitted curves for all the features in the dataset. The horizontal dotted line displays SHAP values at zero (y = 0), whereas the vertical blue dotted line shows where the feature value intersects it. This intersection signifies the threshold at which the feature’s contribution to the prediction shifts from negative to positive or vice versa. For instance, for XB, values less than 1.19 negatively impact the model’s predictions, whereas values greater than 1.19 positively influence the predictions. There are two intersections in the case of E, at 10.29 and 41.66. The predictions are negatively affected by values less than and between 10.29 and 41.66, and positively affected by values above this range. However, the exact intersection values are difficult to discern in Figure 13, which prompted the use of Figure 14, which provides a clearer depiction of these critical points. For the SB, HB, TB, and Pf features, the intersections are at 1.11, 2.47, 1.29, and 1.28, respectively. This study developed a novel rock fragmentation prediction model and subsequently a user-friendly web interface through cloud deployment. The web interface overcomes the limitations of requiring software downloads on computers or mobile devices, offering seamless accessibility for rock fragmentation prediction applications. The following is the link to the web application accessed on 15 October 2024: https://rockfragmentation.streamlit.app/Data_statistics.

6. Conclusions

The prediction of MFS using AI-based techniques has practical implications in the mining industry, since it improves environmental and miners’ safety while lowering costs. To model MFS prediction, this study employed three AI tree-based techniques coupled with the TPE optimization algorithm. Furthermore, this study utilized the SHAP technique to systematically evaluate the stability, robustness, and interpretability of the model. Lastly, an interactive web application was developed to facilitate the MFS predictions. The main conclusions of this study are as follows:
(1)
Adding 3% noise to augment the dataset did not significantly distort the original dataset. Moreover, the large-scale database of 3740 samples provided deeper insights into the input and output parameters and thus enhanced the model’s predictive capabilities;
(2)
The model evaluation results demonstrated that the TPE-ET model performed better than the other models in predicting MFS, achieving R2, RMSE, MAE, and max error optimal values of 0.93, 0.04, 0.03, and 0.25 on the testing set;
(3)
The model interpretability results illustrated that rock parameters and geological conditions were the most significant parameters in predicting MFS. In this study, XB (m) and E (GPa) had the most significant impact and a positive contribution to the models’ predictions.
The tree-based algorithms utilized in this study produced outstanding predictions. However, there are still limitations, such as (1) the fact that this study used a small number of input parameters, which can limit the true precision of the model’s prediction considering that there are many factors that influence MFS; (2) the fact that other powerful tree-based algorithms such as the histogram-based gradient boosting algorithm, extreme gradient boosting algorithm, and gradient boosting algorithm were not explored in the current study; and (3) the fact that noise addition during data augmentation reduced the data quality. Some of the recommendations for future research include the following: (1) employing ensemble methods may enhance the generalization capacities of the models, for example integrating tree-based models with other methods such as SVMs or ANNs; (2) integrating physics-informed models with machine learning models would enhance the accuracy and reliability of this method, as machine learning models do not adhere to the physical laws governing the time-dependent dynamics of fragmentation systems; and (3) utilizing sensors to gather real-time fragmentation data and employing machine learning models for instantaneous predictions.

Author Contributions

Conceptualization: M.M., S.H. and C.L.; Methodology: M.M., C.L. and J.Z.; Investigation: M.M. and J.Z.; Writing—original draft preparation: M.M. and S.H.; Writing—review and editing: S.H., C.L. and J.Z.; Visualization: M.M., C.L. and J.Z.; Funding acquisition: J.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research is partially supported by the Distinguished Youth Science Foundation of Hunan Province of China (2022JJ10073).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used in this study are from published research: Sharma et al. [55] (Sharma, S.K. and Rai, P. Establishment of blasting design parameters influencing mean fragment size using state-of-art statistical tools and techniques. Measurement, 2017, 96: 34–51); Hudaverdi et al. [57] (Hudaverdi, T., Kulatilake, P. and Kuzu, C. Prediction of blast fragmentation using multivariate analysis procedures. Inter-national Journal for Numerical and Analytical Methods in Geomechanics, 2011, 35: 1318–1333); Renchao and Pinguang [58] (Renchao, W. and Pinguang, Z. Study on blasting fragmentation prediction model based on random forest regression method. Journal of Hydropower, 2020: 23–34.).

Acknowledgments

The authors want to thank all the members who give us lots of help and co-operation.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Hu, H.; Lu, W.; Yan, P.; Chen, M.; Gao, Q.; Yang, Z. A new horizontal rock dam foundation blasting technique with a shock-reflection device arranged at the bottom of vertical borehole. Eur. J. Environ. Civ. Eng. 2020, 24, 481–499. [Google Scholar] [CrossRef]
  2. Zhou, J.; Chen, C.; Du, K.; Armaghani, D.J.; Li, C. A new hybrid model of information entropy and unascertained measurement with different membership functions for evaluating destressability in burst-prone underground mines. Eng. Comput. 2022, 38, 381–399. [Google Scholar] [CrossRef]
  3. Li, C.; Zhou, J. Prediction and optimization of adverse responses for a highway tunnel after blasting excavation using a novel hybrid multi-objective intelligent model. Transp. Geotech. 2024, 45, 101228. [Google Scholar] [CrossRef]
  4. Yang, Z.; He, B.; Liu, Y.; Wang, D.; Zhu, G. Classification of rock fragments produced by tunnel boring machine using convolutional neural networks. Autom. Constr. 2021, 125, 103612. [Google Scholar] [CrossRef]
  5. Ebrahimi, E.; Monjezi, M.; Khalesi, M.R.; Armaghani, D.J. Prediction and optimization of back-break and rock fragmentation using an artificial neural network and a bee colony algorithm. Bull. Eng. Geol. Environ. 2016, 75, 27–36. [Google Scholar] [CrossRef]
  6. Khandelwal, M.; Monjezi, M. Prediction of Backbreak in Open-Pit Blasting Operations Using the Machine Learning Method. Rock Mech. Rock Eng. 2013, 46, 389–396. [Google Scholar] [CrossRef]
  7. Monjezi, M.; Khoshalan, H.A.; Varjani, A.Y. Prediction of flyrock and backbreak in open pit blasting operation: A neuro-genetic approach. Arab. J. Geosci. 2012, 5, 441–448. [Google Scholar] [CrossRef]
  8. Esmaeili, M.; Osanloo, M.; Rashidinejad, F.; Bazzazi, A.A.; Taji, M. Multiple regression, ANN and ANFIS models for prediction of backbreak in the open pit blasting. Eng. Comput. 2014, 30, 549–558. [Google Scholar] [CrossRef]
  9. Zhou, J.; Asteris, P.G.; Armaghani, D.J.; Pham, B.T. Prediction of ground vibration induced by blasting operations through the use of the Bayesian Network and random forest models. Soil Dyn. Earthq. Eng. 2020, 139, 106390. [Google Scholar] [CrossRef]
  10. Armaghani, D.J.; Hajihassani, M.; Mohamad, E.T.; Marto, A.; Noorani, S.A. Blasting-induced flyrock and ground vibration prediction through an expert artificial neural network based on particle swarm optimization. Arab. J. Geosci. 2014, 7, 5383–5396. [Google Scholar] [CrossRef]
  11. Görgülü, K.; Arpaz, E.; Demirci, A.; Koçaslan, A.; Dilmaç, M.K.; Yüksek, A.G. Investigation of blast-induced ground vibrations in the Tülü boron open pit mine. Bull. Eng. Geol. Environ. 2013, 72, 555–564. [Google Scholar] [CrossRef]
  12. Raina, A.; Murthy, V.; Soni, A. Flyrock in bench blasting: A comprehensive review. Bull. Eng. Geol. Environ. 2014, 73, 1199–1209. [Google Scholar] [CrossRef]
  13. Jang, H.; Kitahara, I.; Kawamura, Y.; Endo, Y.; Topal, E.; Degawa, R.; Mazara, S. Development of 3D rock fragmentation measurement system using photogrammetry. Int. J. Min. Reclam. Environ. 2020, 34, 294–305. [Google Scholar] [CrossRef]
  14. Ouchterlony, F.; Sanchidrián, J.A. A review of the development of better prediction equations for blast fragmentation. Rock Dyn. Appl. 3 2018, 11, 25–45. [Google Scholar] [CrossRef]
  15. Zhang, Z.-X.; Hou, D.-F.; Guo, Z.; He, Z.; Zhang, Q. Experimental study of surface constraint effect on rock fragmentation by blasting. Int. J. Rock Mech. Min. Sci. 2020, 128, 104278. [Google Scholar] [CrossRef]
  16. Raina, A.K.; Vajre, R.; Sangode, A.; Chandar, K.R. Application of artificial intelligence in predicting rock fragmentation: A review. Appl. Artif. Intell. Min. Geotech. Geoengin. 2024, 291–314. [Google Scholar] [CrossRef]
  17. Li, E.; Yang, F.; Ren, M.; Zhang, X.; Zhou, J.; Khandelwal, M. Prediction of blasting mean fragment size using support vector regression combined with five optimization algorithms. J. Rock Mech. Geotech. Eng. 2021, 13, 1380–1397. [Google Scholar] [CrossRef]
  18. Kuznetsov, V. The mean diameter of the fragments formed by blasting rock. Sov. Min. Sci. 1973, 9, 144–148. [Google Scholar] [CrossRef]
  19. Cunningham, C. The Kuz-Ram model for prediction of fragmentation from blasting. In First International Symposium on Rock Fragmentation by Blasting; Luleå University of Technology: Luleå, Sweden, 1983. [Google Scholar]
  20. Cunningham, C. The Kuz-Ram fragmentation model–20 years on. In Brighton Conference Proceedings; European Federation of Explosives Engineer: Brighton, UK, 2005. [Google Scholar]
  21. Adebola, J.M.; Ajayi, O.D.; Elijah, P. Rock fragmentation prediction using Kuz-Ram model. J. Environ. Earth Sci. 2016, 6, 110–115. [Google Scholar]
  22. Ouchterlony, F. The Swebrec© function: Linking fragmentation by blasting and crushing. Min. Technol. 2005, 114, 29–44. [Google Scholar] [CrossRef]
  23. Sanchidrián, J.A.; Ouchterlony, F. A distribution-free description of fragmentation by blasting based on dimensional analysis. Rock Mech. Rock Eng. 2017, 50, 781–806. [Google Scholar] [CrossRef]
  24. Ouchterlony, F. Influence of Blasting on the Size Distribution and Properties of Muckpile Fagments: A State-of-the-Art Review; Luleå University of Technology: Luleå, Sweden, 2003. [Google Scholar]
  25. Gheibie, S.; Aghababaei, H.; Hoseinie, S.; Pourrahimian, Y. Modified Kuz—Ram fragmentation model and its use at the Sungun Copper Mine. Int. J. Rock Mech. Min. Sci. 2009, 46, 967–973. [Google Scholar] [CrossRef]
  26. Bergmann, O.R.; Riggle, J.W.; Wu, F.C. Model rock blasting—Effect of explosives properties and other variables on blasting results. Int. J. Rock Mech. Min. Sci. Geomech. Abstr. 1973, 10, 585–612. [Google Scholar] [CrossRef]
  27. Kumar, M.; Kumar, V.; Biswas, R.; Samui, P.; Kaloop, M.R.; Alzara, M.; Yosri, A.M. Hybrid ELM and MARS-based prediction model for bearing capacity of shallow foundation. Processes 2022, 10, 1013. [Google Scholar] [CrossRef]
  28. Zhou, J.; Li, E.; Yang, S.; Wang, M.; Shi, X.; Yao, S.; Mitri, H.S. Slope stability prediction for circular mode failure using gradient boosting machine approach based on an updated database of case histories. Saf. Sci. 2019, 118, 505–518. [Google Scholar] [CrossRef]
  29. Zhou, J.; Huang, S.; Wang, M.; Qiu, Y. Performance evaluation of hybrid GA–SVM and GWO–SVM models to predict earthquake-induced liquefaction potential of soil: A multi-dataset investigation. Eng. Comput. 2022, 38, 4197–4215. [Google Scholar] [CrossRef]
  30. Biswas, R.; Li, E.; Zhang, N.; Kumar, S.; Rai, B.; Zhou, J. Development of hybrid models using metaheuristic optimization techniques to predict the carbonation depth of fly ash concrete. Constr. Build. Mater. 2022, 346, 128483. [Google Scholar] [CrossRef]
  31. Shen, Y.; Wu, S.; Wang, Y.; Wang, J.; Yang, Z. Interpretable model for rockburst intensity prediction based on Shapley values-based Optuna-random forest. Undergr. Space 2024, 21, 198–214. [Google Scholar] [CrossRef]
  32. Mame, M.; Qiu, Y.; Huang, S.; Du, K.; Zhou, J. Mean Block Size Prediction in Rock Blast Fragmentation Using TPE-Tree-Based Model Approach with SHapley Additive exPlanations. Min. Metall. Explor. 2024, 41, 2325–2340. [Google Scholar] [CrossRef]
  33. Yari, M.; He, B.; Armaghani, D.J.; Abbasi, P.; Mohamad, E.T. A novel ensemble machine learning model to predict mine blasting–induced rock fragmentation. Bull. Eng. Geol. Environ. 2023, 82, 187. [Google Scholar] [CrossRef]
  34. Shi, X.-Z.; Zhou, J.; Wu, B.-B.; Huang, D.; Wei, W. Support vector machines approach to mean particle size of rock fragmentation due to bench blasting prediction. Trans. Nonferrous Met. Soc. China 2012, 22, 432–441. [Google Scholar] [CrossRef]
  35. Monjezi, M.; Mohamadi, H.A.; Barati, B.; Khandelwal, M. Application of soft computing in predicting rock fragmentation to reduce environmental blasting side effects. Arab. J. Geosci. 2014, 7, 505–511. [Google Scholar] [CrossRef]
  36. Dimitraki, L.; Christaras, B.; Marinos, V.; Vlahavas, I.; Arampelos, N. Predicting the average size of blasted rocks in aggregate quarries using artificial neural networks. Bull. Eng. Geol. Environ. 2019, 78, 2717–2729. [Google Scholar] [CrossRef]
  37. Kulatilake, P.H.S.W.; Hudaverdi, T.; Wu, Q. New Prediction Models for Mean Particle Size in Rock Blast Fragmentation. Geotech. Geol. Eng. 2012, 30, 665–684. [Google Scholar] [CrossRef]
  38. Kulatilake, P.H.S.W.; Qiong, W.; Hudaverdi, T.; Kuzu, C. Mean particle size prediction in rock blast fragmentation using neural networks. Eng. Geol. 2010, 114, 298–311. [Google Scholar] [CrossRef]
  39. Shams, S.; Monjezi, M.; Majd, V.J.; Armaghani, D.J. Application of fuzzy inference system for prediction of rock fragmentation induced by blasting. Arab. J. Geosci. 2015, 8, 10819–10832. [Google Scholar] [CrossRef]
  40. Ghaeini, N.; Mousakhani, M.; Amnieh, H.B.; Jafari, A. Prediction of blasting-induced fragmentation in Meydook copper mine using empirical, statistical, and mutual information models. Arab. J. Geosci. 2017, 10, 409. [Google Scholar] [CrossRef]
  41. Asl, P.F.; Monjezi, M.; Hamidi, J.K.; Armaghani, D.J. Optimization of flyrock and rock fragmentation in the Tajareh limestone mine using metaheuristics method of firefly algorithm. Eng. Comput. 2018, 34, 241–251. [Google Scholar] [CrossRef]
  42. Gao, W.; Karbasi, M.; Hasanipanah, M.; Zhang, X.; Guo, J. Developing GPR model for forecasting the rock fragmentation in surface mines. Eng. Comput. 2018, 34, 339–345. [Google Scholar] [CrossRef]
  43. Sayevand, K.; Arab, H.; Golzar, S.B. Development of imperialist competitive algorithm in predicting the particle size distribution after mine blasting. Eng. Comput. 2018, 34, 329–338. [Google Scholar] [CrossRef]
  44. Hasanipanah, M.; Amnieh, H.B.; Arab, H.; Zamzam, M.S. Feasibility of PSO–ANFIS model to estimate rock fragmentation produced by mine blasting. Neural Comput. Appl. 2018, 30, 1015–1024. [Google Scholar] [CrossRef]
  45. Zhou, J.; Li, C.; Arslan, C.A.; Hasanipanah, M.; Amnieh, H.B. Performance evaluation of hybrid FFA-ANFIS and GA-ANFIS models to predict particle size distribution of a muck-pile after blasting. Eng. Comput. 2021, 37, 265–274. [Google Scholar] [CrossRef]
  46. Huang, J.; Asteris, P.G.; Pasha, S.M.K.; Mohammed, A.S.; Hasanipanah, M. A new auto-tuning model for predicting the rock fragmentation: A cat swarm optimization algorithm. Eng. Comput. 2022, 38, 2209–2220. [Google Scholar] [CrossRef]
  47. Fang, Q.; Nguyen, H.; Bui, X.-N.; Nguyen-Thoi, T.; Zhou, J. Modeling of rock fragmentation by firefly optimization algorithm and boosted generalized additive model. Neural Comput. Appl. 2021, 33, 3503–3519. [Google Scholar] [CrossRef]
  48. Zhang, S.; Bui, X.-N.; Trung, N.-T.; Nguyen, H.; Bui, H.-B. Prediction of rock size distribution in mine bench blasting using a novel ant colony optimization-based boosted regression tree technique. Nat. Resour. Res. 2020, 29, 867–886. [Google Scholar] [CrossRef]
  49. Amoako, R.; Jha, A.; Zhong, S. Rock fragmentation prediction using an artificial neural network and support vector regression hybrid approach. Mining 2022, 2, 233–247. [Google Scholar] [CrossRef]
  50. Mehrdanesh, A.; Monjezi, M.; Khandelwal, M.; Bayat, P. Application of various robust techniques to study and evaluate the role of effective parameters on rock fragmentation. Eng. Comput. 2023, 39, 1317–1327. [Google Scholar] [CrossRef]
  51. Li, E.; Zhou, J.; Biswas, R.; Ahmed, Z.E.M. Fragmentation by blasting size prediction using SVR-GOA and SVR-KHA techniques. In Applications of Artificial Intelligence in Mining, Geotechnical and Geoengineering; Elsevier: Amsterdam, The Netherlands, 2024; pp. 343–360. [Google Scholar]
  52. Rong, K.; Xu, X.; Wang, H.; Yang, J. Prediction of the mean fragment size in mine blasting operations by deep learning and grey wolf optimization algorithm. Earth Sci. Inform. 2024, 17, 2903–2919. [Google Scholar] [CrossRef]
  53. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  54. Sachpazis, C. Correlating Schmidt hardness with compressive strength and Young’s modulus of carbonate rocks. Bull. Eng. Geol. Environ. 1990, 42, 75–83. [Google Scholar] [CrossRef]
  55. Sharma, S.K.; Rai, P. Establishment of blasting design parameters influencing mean fragment size using state-of-art statistical tools and techniques. Measurement 2017, 96, 34–51. [Google Scholar] [CrossRef]
  56. Ouchterlony, F.; Niklasson, B.; Abrahamsson, S. Fragmentation monitoring of production blasts at MRICA. In International Symposium on Rock Fragmentation by Blasting: 26/08/1990–31/08/1990; The Australian Institute of Mining and Metallurgy: Carlton, Australia, 1990. [Google Scholar]
  57. Hudaverdi, T.; Kulatilake, P.; Kuzu, C. Prediction of blast fragmentation using multivariate analysis procedures. Int. J. Numer. Anal. Methods Geomech. 2011, 35, 1318–1333. [Google Scholar] [CrossRef]
  58. Renchao, W.; Pinguang, Z. Study on blasting fragmentation prediction model based on random forest regression method. J. Hydropower 2020, 39, 89–101. [Google Scholar]
  59. Mikołajczyk, A.; Grochowski, M. Data augmentation for improving deep learning in image classification problem. In Proceedings of the International Interdisciplinary PhD Workshop (IIPhDW), Swinoujscie, Poland, 9–12 May 2018; IEEE: New York, NY, USA, 2018. [Google Scholar]
  60. Iwana, B.K.; Uchida, S. An empirical survey of data augmentation for time series classification with neural networks. PLoS ONE 2021, 16, e0254841. [Google Scholar] [CrossRef] [PubMed]
  61. Dong, H.; Wang, J.; Wu, X.; Zhou, M.; Lü, J. Gaussian noise data augmentation-based delay prediction for high-speed railways. IEEE Intell. Transp. Syst. Mag. 2023, 15, 8–18. [Google Scholar] [CrossRef]
  62. Xi, B.; Li, E.; Fissha, Y.; Zhou, J.; Segarra, P. LGBM-based modeling scenarios to compressive strength of recycled aggregate concrete with SHAP analysis. Mech. Adv. Mater. Struct. 2023, 31, 5999–6014. [Google Scholar] [CrossRef]
  63. Zhang, W.; Lee, D.; Lee, J.; Lee, C. Residual strength of concrete subjected to fatigue based on machine learning technique. Struct. Concr. 2022, 23, 2274–2287. [Google Scholar] [CrossRef]
  64. Wahba, M.; Essam, R.; El-Rawy, M.; Al-Arifi, N.; Abdalla, F.; Elsadek, W.M. Forecasting of flash flood susceptibility mapping using random forest regression model and geographic information systems. Heliyon 2024, 10, e33982. [Google Scholar] [CrossRef]
  65. Breiman, L. Classification and Regression Trees; Routledge: Oxfordshire, UK, 2017. [Google Scholar]
  66. Breiman, L. Bagging predictors. Mach. Learn. 1996, 24, 123–140. [Google Scholar] [CrossRef]
  67. Geurts, P.; Ernst, D.; Wehenkel, L. Extremely randomized trees. Mach. Learn. 2006, 63, 3–42. [Google Scholar] [CrossRef]
  68. Mishra, G.; Sehgal, D.; Valadi, J.K. Quantitative structure activity relationship study of the anti-hepatitis peptides employing random forests and extra-trees regressors. Bioinformation 2017, 13, 60. [Google Scholar] [CrossRef]
  69. John, V.; Liu, Z.; Guo, C.; Mita, S.; Kidono, K. Real-time lane estimation using deep features and extra trees regression. In Image and Video Technology, Proceedings of the 7th Pacific-Rim Symposium, PSIVT 2015, Auckland, New Zealand, 25–27 November 2015, Revised Selected Papers 7; Springer: Berlin/Heidelberg, Germany, 2016. [Google Scholar]
  70. Dai, L.; Feng, D.; Pan, Y.; Wang, A.; Ma, Y.; Xiao, Y.; Zhang, J. Quantitative principles of dynamic interaction between rock support and surrounding rock in rockburst roadways. Int. J. Min. Sci. Technol. 2025, 35, 41–55. [Google Scholar] [CrossRef]
  71. Zhang, Y.; Haghani, A. A gradient boosting method to improve travel time prediction. Transp. Res. Part C: Emerg. Technol. 2015, 58, 308–324. [Google Scholar] [CrossRef]
  72. Friedman, J.H. Greedy function approximation: A gradient boosting machine. Ann. Stat. 2001, 29, 1189–1232. [Google Scholar] [CrossRef]
  73. Schapire, R.E. The strength of weak learnability. Mach. Learn. 1990, 5, 197–227. [Google Scholar] [CrossRef]
  74. Natekin, A.; Knoll, A. Gradient boosting machines, a tutorial. Front. Neurorobotics 2013, 7, 21. [Google Scholar] [CrossRef]
  75. Bergstra, J.; Bardenet, R.; Bengio, Y.; Kégl, B. Algorithms for Hyper-Parameter Optimization, 2011. J. Phys. Conf. Ser. 2021. [Google Scholar]
  76. Rong, G.; Li, K.; Su, Y.; Tong, Z.; Liu, X.; Zhang, J.; Zhang, Y.; Li, T. Comparison of tree-structured parzen estimator optimization in three typical neural network models for landslide susceptibility assessment. Remote Sens. 2021, 13, 4694. [Google Scholar] [CrossRef]
  77. Zhang, W.; Wu, C.; Zhong, H.; Li, Y.; Wang, L. Prediction of undrained shear strength using extreme gradient boosting and random forest based on Bayesian optimization. Geosci. Front. 2021, 12, 469–477. [Google Scholar] [CrossRef]
  78. Shen, K.; Qin, H.; Zhou, J.; Liu, G. Runoff probability prediction model based on natural Gradient boosting with tree-structured parzen estimator optimization. Water 2022, 14, 545. [Google Scholar] [CrossRef]
  79. Ekanayake, I.; Meddage, D.; Rathnayake, U. A novel approach to explain the black-box nature of machine learning in compressive strength predictions of concrete using Shapley additive explanations (SHAP). Case Stud. Constr. Mater. 2022, 16, e01059. [Google Scholar] [CrossRef]
  80. Zhang, C.; Cho, S.; Vasarhelyi, M. Explainable Artificial Intelligence (XAI) in auditing. Int. J. Account. Inf. Syst. 2022, 46, 100572. [Google Scholar] [CrossRef]
  81. Kim, Y.; Kim, Y. Explainable heat-related mortality with random forest and SHapley Additive exPlanations (SHAP) models. Sustain. Cities Soc. 2022, 79, 103677. [Google Scholar] [CrossRef]
  82. Ngo, A.Q.; Nguyen, L.Q.; Tran, V.Q. Developing interpretable machine learning-Shapley additive explanations model for unconfined compressive strength of cohesive soils stabilized with geopolymer. PLoS ONE 2023, 18, e0286950. [Google Scholar] [CrossRef] [PubMed]
  83. Lundberg, S.M.; Lee, S.-I. A unified approach to interpreting model predictions. Adv. Neural Inf. Process. Syst. 2017, 30. [Google Scholar]
Figure 1. Correlation plot of the large-scale dataset.
Figure 1. Correlation plot of the large-scale dataset.
Applsci 15 08363 g001
Figure 2. The framework of the RF algorithm.
Figure 2. The framework of the RF algorithm.
Applsci 15 08363 g002
Figure 3. The framework of the ET algorithm.
Figure 3. The framework of the ET algorithm.
Applsci 15 08363 g003
Figure 4. The framework of the three proposed models.
Figure 4. The framework of the three proposed models.
Applsci 15 08363 g004
Figure 6. Hyperparameter importances of each model.
Figure 6. Hyperparameter importances of each model.
Applsci 15 08363 g006
Figure 7. Comparison of the models during the optimization process.
Figure 7. Comparison of the models during the optimization process.
Applsci 15 08363 g007
Figure 8. The performance score ranking results of the models: (a) training phase and (b) testing phase.
Figure 8. The performance score ranking results of the models: (a) training phase and (b) testing phase.
Applsci 15 08363 g008
Figure 9. Scatter plots of the models during the training phase: (a) TPE-GB model, (b) TPE-ET model, and (c) TPE-RF model.
Figure 9. Scatter plots of the models during the training phase: (a) TPE-GB model, (b) TPE-ET model, and (c) TPE-RF model.
Applsci 15 08363 g009
Figure 10. Scatter plots of the models during the testing phase: (a) TPE-ET model, (b) TPE-GB model, and (c) TPE-RF model.
Figure 10. Scatter plots of the models during the testing phase: (a) TPE-ET model, (b) TPE-GB model, and (c) TPE-RF model.
Applsci 15 08363 g010
Figure 11. A bee swarm plot of the SHAP interaction matrix.
Figure 11. A bee swarm plot of the SHAP interaction matrix.
Applsci 15 08363 g011
Figure 12. A combination of a local bar plot and a bee swarm plot of feature importances.
Figure 12. A combination of a local bar plot and a bee swarm plot of feature importances.
Applsci 15 08363 g012
Figure 13. LOWESS fit curve with SHAP scatter plot of each feature.
Figure 13. LOWESS fit curve with SHAP scatter plot of each feature.
Applsci 15 08363 g013
Figure 14. A scatter plot of BD and SHAP values with a LOWESS fit curve.
Figure 14. A scatter plot of BD and SHAP values with a LOWESS fit curve.
Applsci 15 08363 g014
Table 1. AI-based techniques utilized over the years to predict rock fragmentation.
Table 1. AI-based techniques utilized over the years to predict rock fragmentation.
SourceMethods Input ParametersOutput ParameterNumber of DatasetsPerformance
[17]GWO-v-SVRD, H, J, S, B, ST, H/B, J/B, B/D, L/Wd, NH, L, Wd, S/B, ST/B, De, Qe, PF, UCSX5076R2 = 0.8353
[32]TPE-ETS/B, H/B, B/D, ST/B, PF, E, XBX50103R2 = 0.9463
[34]SVMS/B, H/B, B/D, ST/B, PF, E, XBXm102R2 = 0.962
[35]ANNB, S, PF, NR, D, MC, ST, HX50135R2 = 0.941
[36]ANNBI, PF, QBX50100R2 = 0.8
[37]ANNS/B, HL/B, B/D, ST/B, PF, XB, EXm109R2 = 0.94
[38]BPNNS/B, HL/B, B/D, ST/B, PF, XB, EXm91R2 = 0.941
[39]FISB, S, D, Sch, DJ, PF, STX80185R2 = 0.922
[40]MIUCS, P, RQD, JS, ρ, q, B, ST, S/D, JPOX8036R2 = 0.81
[41]ANNB, S, HL, SD, ST, MC, PF, GSIX80200R2 = 0.94
[42]GPRB, S, ST, PF, MCX8072R2 = 0.948
[43]ICAMC, B, S, ST, PF, RMRX8080R2 = 0.947
[44]PSO-ANFISB, S, ST, q, MCX8072R2 = 0.89
[45]FFA-ANFISB, S, ST, PF, MCX8072R2 = 0.98
[46]CSOq, B, RMR, MC, ST, S X8075R2 = 0.985
[45]GA-ANFISB, S, ST, PF, MC, RMRX8088R2 = 0.989
[47]FFA-BGAMPF, MC, S, ST, B, H X100136R2 = 0.98
[47]FFA-BGAMW, P, H, T, S, B, SDR136R2 = 0.98
[48]ACO-BRTPF, MC, S, ST, B, H X100136R2 = 0.962
[49]ANNS/B, H/B, B/D, ST/B, PF, E, XBX50102R2 = 0.87
[50]ANNB, S, H, D, T, PF, Is50, UCS, UTS, ρ, E, Vp, SHV, U, RQD, C, φ, XBX50353R2 = 0.986
[51]GOA-SVRD, H, J, S, B, ST, H/B, J/B, B/D, L/Wd, NH, L, Wd, S/B, ST/B, De, Qe, PF, UCSX5076R2 = 85.83
[52]GWO-CNNS/B, H/B, B/D, ST/B, PF, E, XBX504540R2 = 0.89772
Note: SVM—support vector machines; FIS—fuzzy inference system; MI—mutual information; GPR—Gaussian process regressor; ICA—imperialist competitive algorithm; PSO—particle swarm optimization; ANFIS—adaptive neural-fuzzy inference system; FFA—firefly algorithm; GA—genetic algorithm; CSO—cat swarm optimization; ACO—ant colony optimization; BRT—boosted regression tree; BGAM—boosted generalized additive model; GWO—grey wolf optimization; GOA—grasshopper optimization algorithm; D—hole diameter (mm); H—bench height (m); S—spacing (m); B—burden; PF—powder factor (kg/m3); ST—stemming; E—elastic modulus (GPa); SDR—size of distributed rock (m); UCS—uniaxial compressive strength (MPa); UTS—uniaxial tensile strength (MPa); RQD—rock quality designation (%); Sch—Schmidt hammer rebound number; Qe—total explosive amount (t); J—sub-grade drilling (m); L—length (m); NH—number of holes; Wd—width (m); RMR—rock mass rating; XB—in situ block size (m); Xm—mean particle size (cm); X80—80% passing size (cm); X100—100% passing size (cm); X50—50% passing size (cm). φ—friction angle; ρ—density, C—cohesion; SHV—Schmidt hardness value; JPO—joint plane orientation/bench face ratio; Vp—P wave velocity; JS—joint spacing (m); ρ—rock density (t/m3); q—specific charge (kg/m3).; Is50—point load strength index (MPa); CNN—convolutional neural network; SVR—support vector regression; MC—charge per delay (kg/ms); GSI—Geological Strength Index; QB—quantity of blasted rock pile (t).
Table 2. Data collected from the published literature.
Table 2. Data collected from the published literature.
Data SourceBlast SamplesInput ParametersOutput Parameters
[55]76D, H, J, S, B, ST, L, Wd, S.B, T.B, H.B, J.B, B.D, L.W, NH, Qe, De, PF, UCSX50 (m)
[57]103SB, HB, BD, TB, Pf(kg/m3), XB(m), EX50 (m)
[58]8SB, HB, BD, TB, Pf(kg/m3), XB(m), E X50 (m)
Total187
Table 3. Large-scale data summary statistics.
Table 3. Large-scale data summary statistics.
Parameters Min. ValueMax. ValueMeanStandard Deviation
S/B0.92671.79211.17880.1093
H/B1.24986.86833.21531.3955
B/D17.940852.224229.34554.7601
T/B0.43534.75131.04770.5782
PF (kg/m3)0.16252.57171.02590.6302
XB (m)0.03072.87241.20290.4764
E (GPa)8.833460.095723.779016.2551
X50 (m)0.01840.99300.31750.1574
Table 5. Evaluation indices of models during the training phase.
Table 5. Evaluation indices of models during the training phase.
ModelsR2RMSEMAEMax ErrorScores
TPE-ET0.970.030.020.14
Rank11114
TPE-GB0.970.030.020.11
Rank11136
TPE-RF0.970.030.020.12
Rank11125
Table 6. Evaluation indices of models during the testing phase.
Table 6. Evaluation indices of models during the testing phase.
ModelsR2RMSEMAEMax ErrorScores
TPE-ET0.930.040.030.25
Rank11136
TPE-GB0.920.040.030.28
Rank21115
TPE-RF0.920.040.030.26
Rank21126
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Mame, M.; Huang, S.; Li, C.; Zhou, J. Application of Extra-Trees Regression and Tree-Structured Parzen Estimators Optimization Algorithm to Predict Blast-Induced Mean Fragmentation Size in Open-Pit Mines. Appl. Sci. 2025, 15, 8363. https://doi.org/10.3390/app15158363

AMA Style

Mame M, Huang S, Li C, Zhou J. Application of Extra-Trees Regression and Tree-Structured Parzen Estimators Optimization Algorithm to Predict Blast-Induced Mean Fragmentation Size in Open-Pit Mines. Applied Sciences. 2025; 15(15):8363. https://doi.org/10.3390/app15158363

Chicago/Turabian Style

Mame, Madalitso, Shuai Huang, Chuanqi Li, and Jian Zhou. 2025. "Application of Extra-Trees Regression and Tree-Structured Parzen Estimators Optimization Algorithm to Predict Blast-Induced Mean Fragmentation Size in Open-Pit Mines" Applied Sciences 15, no. 15: 8363. https://doi.org/10.3390/app15158363

APA Style

Mame, M., Huang, S., Li, C., & Zhou, J. (2025). Application of Extra-Trees Regression and Tree-Structured Parzen Estimators Optimization Algorithm to Predict Blast-Induced Mean Fragmentation Size in Open-Pit Mines. Applied Sciences, 15(15), 8363. https://doi.org/10.3390/app15158363

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop