Next Article in Journal
Activation Energy Performance through Magnetized Hybrid Fe3O4PP Nanofluids Flow with Impact of the Cluster Interfacial Nanolayer
Next Article in Special Issue
Risk Evaluation Model of Coal Spontaneous Combustion Based on AEM-AHP-LSTM
Previous Article in Journal
CNN-Hyperparameter Optimization for Diabetic Maculopathy Diagnosis in Optical Coherence Tomography and Fundus Retinography
Previous Article in Special Issue
SEIARN: Intelligent Early Warning Model of Epidemic Spread Based on LSTM Trajectory Prediction
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Rock Burst Intensity Classification Prediction Model Based on a Bayesian Hyperparameter Optimization Support Vector Machine

1
College of Mining Engineering, North China University of Science and Technology, Tangshan 063210, China
2
College of Science, North China University of Science and Technology, Tangshan 063210, China
*
Author to whom correspondence should be addressed.
Mathematics 2022, 10(18), 3276; https://doi.org/10.3390/math10183276
Submission received: 2 August 2022 / Revised: 14 August 2022 / Accepted: 5 September 2022 / Published: 9 September 2022
(This article belongs to the Special Issue Engineering Calculation and Data Modeling)

Abstract

:
Rock burst disasters occurring in underground high-stress rock mass mining and excavation engineering seriously threaten the safety of workers and hinders the progress of engineering construction. Rock burst classification prediction is the basis of reducing and even eliminating rock burst hazards. Currently, most of mainstream discriminant models for rock burst grade prediction are based on small samples. Comprehensive selection according to many pieces of literature, the maximum tangential stress of surrounding rock and rock uniaxial compressive strength ratio coefficient (stress state parameter), rock uniaxial compressive strength and uniaxial tensile strength ratio (brittleness modulus), and the elastic energy index are used as a grading evaluation index of rock burst based on the collection of different construction engineering instances of rock burst in 114 groups of extensive sample data in different regions of the world, which are used to carry out the training study. The representativeness and accuracy of the index selection were verified by the indicator variance analysis and Spearman correlation coefficient hypothesis test. The Intelligent Rock burst Identification System (IRIS) based on an optimizable SVM model was established using data set samples. After extensive data cross-validation training, the accuracy of the SVM discriminant analysis model can reach 95.6%, which is significantly better than the prediction accuracy of the traditional SVM model of 71.9%. The model is used to classify and predict the rock burst intensity of 10 typical projects at home and abroad. The prediction results are consistent with the actual rock burst intensity, which is better than the discriminant model based on small sample data and other existing prediction models. The application of engineering examples shows that the results of the rock burst intensity classification prediction model based on extensive sample data processing analysis and the SVM discriminant method are in good agreement with the actual rock burst intensity, which can effectively provide a reference for the prediction of rock burst intensity grade in a construction area.

1. Introduction

Rock burst is a failure phenomenon in excavating underground rock mass with high stress. Under the action of excavation or other load disturbances, hard and brittle surrounding rocks in a high-stress state rapidly release the accumulated elastic energy, resulting in the dynamic instability disaster of rock spalling, fragmentation and ejection. Rock burst is characterized by its sudden occurrence and the significant harm caused, which not only poses a significant threat to the construction personnel and mechanical equipment, but also delays the construction period, even destroying the whole project and inducing earthquakes [1]. On 28 November 2009, an intense rock burst occurred in the drainage tunnel of Jinping Ii Hydropower Station, resulting in seven deaths and direct economic losses of hundreds of millions of CNY [2]. On 9 March 2014, a rock collapse caused by a rock burst occurred in the Dulong River highway tunnel in Gongshan, Yunnan province, killing three construction workers [3]. With the development of mining activities to a great depth, the increase of buried depth and the increase in stress level, the mining and excavation of underground high-stress rock mass shows a sudden trend of high frequency.
The accurate grading prediction of rock burst intensity has an extremely important theoretical and practical status for the effective prevention and control of rock burst in engineering. In general, rock burst prediction methods can be divided into four categories [4], namely the rock burst criterion method, numerical index method, applied mathematical method and on-site monitoring method. With the continuous development of big data analysis technology, the nonlinear rock burst prediction method (applied mathematical method) that comprehensively considers multiple influencing factors has achieved good rock burst prediction results and is a method with good development prospects in future rock burst intensity prediction. In particular, the typical “training modeling” method can achieve a better prediction effect. The main idea is that a fixed number of rock burst instance samples were trained and taught to establish a prediction model, including the neural network method, support vector machine (SVM) method, Distance discriminant method, Bayes discriminant method, Fisher discriminant method, Gaussian process method, random forest model and many other methods, but some machine learning algorithms have shortcomings in their application process. For example, Naive Bayes [5] has a poor classification effect on sample data sets with a strong index correlation. After testing, the accuracy of the prediction method is stable at 79.8%. The classification effect of random forest on low-dimensional data is not good, and the accuracy of the prediction method is stable at 79.8% after testing. Neural network [5] has a slow convergence speed in the learning process, and the hidden layer neurons need to be selected. After testing, the accuracy of this prediction method is stable at 86.8%. Due to the randomness and complexity of rock burst disasters, it is necessary to select a discriminant model with a good learning effect and high prediction accuracy.
In practical application, the optimized SVM algorithm is better than the traditional SVM algorithm in all aspects, which provides a way to solve the problem of SVM being difficult to calculate and classify large-scale training samples. The advantage of the optimizable SVM algorithm in the training process lies in that it can optimize and adjust the multi-class methods, box constraint level, kernel function, kernel scale and standardized data in the hyperparameter training process in the iterative process until the best classification effect is obtained. Based on the above analysis, the optimized SVM classification prediction model was established and applied to 132 groups of typical rock burst samples for training and testing. The model was compared with XGBoot (the accuracy is stable at 70%), RF (the accuracy is stable at 70%) and SVM models (the accuracy is stable at 80%) in engineering applications, which proved the accuracy, reliability and practicability of the optimized SVM discriminating model (the accuracy is stable at 95.6%).

2. Principle and Method of Rock Burst Intensity Prediction Model

Support Vector Machines (SVM) are a linear classifier model developed in the late 20th century to solve the binary classification problem. They map the feature vectors of the study samples to some points in the space through a specific rule. Their purpose is to draw the distinction line between the two. To distinguish the two kinds of topics accurately and efficiently, even after the training is completed, new issues to be measured are added, and the trained lines can still be classified well. The key to the SVM machine learning algorithm is to select the kernel function and related parameters of the model. Standard kernel functions include the linear kernel function, radial basis kernel function, Laplacian kernel function, and so on. With little subjective influence, this method is one of the most commonly used classifiers with the best effect at present. Since its optimization goal is to minimize structural risk, it has excellent generalization ability, so it has been widely used in the classification and prediction of rock burst intensity. Based on collecting a large number of rock burst samples, the SVM discriminant model for rock burst intensity classification prediction is established and applied in engineering.

2.1. Support Vector Machines

Support Vector Machines (SVM) are a linear classifier model developed in the late 20th century to solve the binary classification problem. They map the feature vectors of the study samples to some points in the space through a specific rule. Its purpose is to draw the distinction line between the two. To distinguish the two kinds of topics accurately and efficiently, even after the training is completed, new issues to be measured are added, and the trained lines can still be classified well. The key to the SVM machine learning algorithm is to select the kernel function and related parameters of the model. Standard kernel functions include the linear kernel function, radial basis kernel function, Laplacian kernel function, and so on. With little subjective influence, this method is one of the most commonly used classifiers with the best effect at present. Since its optimization goal is to minimize structural risk, it has excellent generalization ability, so it has been widely used in the classification and prediction of rock burst intensity. Based on collecting a large number of rock burst samples, the SVM discriminant model for rock burst intensity classification prediction is established and applied in engineering.
(1)
Linear Support Vector Machines. According to the principle of the SVM hyperplane w × x + b = 0   to divide the sample data set, to ensure the maximum classification interval, the hyperplane optimization method is expressed as Formula (1)
m i n ϑ ( w ) = 1 2 | | w | | 2 = 1 2 w T w
Lagrangian functions are usually used to convert samples of initial problems into dual problems, resulting in Formula (2).
m a x μ ( a ) = i = 1 n a i 1 2 i , j = 1 n a i a j y i y j ( x i × x j )
After solving by Formula (2), the optimal classification function can be obtained by Formula (3)
f ( x ) = s g n ( w T + b ) = s g n ( i = 1 n a i * y i ( x i × x ) + b * )
However, in practical application, many data samples are linearly indivisible, so blindly using the linearly inseparable method causes significant errors. To deal with this situation, the problem adaptation formula is changed into Formula (4) after the relaxation variable is introduced.
ϑ ( w ) = 1 2 | | w | | 2 + C i = 1 n ϵ i
where C is the penalty factor.
The Lagrange algorithm was used to solve the problem and Formula (5) was obtained.
m a x μ ( a ) = i = 1 n a i 1 2 i , j = 1 n a i a j y i y j ( x i × x j )
After solving all kinds of coefficients   a i , the classification decision function is obtained, and the Formula (6) is obtained.
f ( x ) = s g n ( i = 1 n a i * y i ( x i × x ) + b * )
(2)
Nonlinear Support Vector Machines. By the theorem of the Mercer condition for any symmetric function K ( x , x ˙ ) , for any φ ( x ) 0 and φ 2 ( x ) d x < , existence K ( x , x ˙ ) φ ( x ) φ ( x ˙ ) d x d x ˙ > 0 was established. The symmetry function   K ( x , x ˙ ) is used to replace the inner product of the optimal classification plane, so the optimization function is Formula (7) (nonlinear data condition).
μ ( a ) = i = 1 n a i 1 2 i , j = 1 n a i a j y i y j ( x i × x j )
The classification decision function is transformed into Formula (8) (nonlinear data condition).
f ( x ) = s g n ( i = 1 n a i * K ( x i × x ) + b * )

2.2. Bayesian Optimization

Based on Bayesian optimization of parameter tuning, in the traditional SVM model, including parameters C and other parts of the above parameters, these parameters and performance have the characteristics of the black-box model, namely between the version of the model and parameters that cannot use the expression to describe, only using the traversal method for construction of the optimal SVM model. Bayesian optimization [6] is an efficient optimization algorithm with sample validity mainly used for parameter tuning machine learning models. This method adopts the Gaussian process to establish a probability proxy model, refers to previous parameter information and iteratively updates prior knowledge, and can determine the optimal parameters of the model in a relatively short time. In general, the expression of the gaussian process is shown in Formula (9).
f ( x ) ~ G P [ m ( x ) , k ( x , x ) ]
In Formula (9), the mean function ( x ) = E ( f ( x ) ) represents the mathematical expectation of sample f ( x ) and the covariance function k ( x , x ) = E { [ f ( x ) m ( x ) ] [ f ( x ) m ( x ) ] } . The gaussian process uses some known points to estimate the mean and variance of the objective function at other points. It then constructs the acquisition function to determine the location of sampling points in the iterative process.

2.3. The Intelligent Rock Burst Identification System

In this paper, Bayesian optimization is adopted as the parameter optimization algorithm of the SVM model. The specific algorithm (based on the Python language environment) flows as follows:
(1)
According to the principle of the SVM model, the value range of equation coefficients such as hyperparameter C is determined and n 0 sampling points are randomly selected. The average accuracy of cross-validation is used as the objective function f , and the parameter combination of the model is used as the independent variable x 0 to construct a surrogate model, and the initial distribution of the objective function and the sampling point set D are obtained.
(2)
Maximize the objective function to obtain the next sampling point x t and the function value f ( x t ) .
(3)
Add sampling points [ x t , f ( x t ) ] to set D , and synchronously update the Gaussian process surrogate model to further fit the distribution of the objective function.
(4)
Return to step (2) and continue the iteration until the algorithm iteration reaches the maximum number of iterations, and then output the optimal sampling point to obtain the optimal hyperparameter C equivalent of the SVM model.

2.4. Great Advantages of Support Vector Machine Based on Bayesian Hyperparameter Optimization

Usually, the influence of a support vector machine is the main problem in building a model to fill in the necessary parameter values; these values are not limited to penalty factor and the parameters of the kernel function are set. Only relying on the existing provisions of researchers and the experience of related papers can lead to a support vector machine (SVM) in which the rock burst prediction accuracy is not huge, and parameter selection and the accuracy value are small. There is no improvement in accuracy; if the parameter selection and the accuracy value are large, it is impossible to judge whether the model result is overfitting. Which parameters to choose and how to choose the parameters are the problems that need to be solved urgently in support vector machine prediction.
The optimizable support vector machine (SVM) model, as a supervised learning algorithm, greatly eliminates the influence of the choice of penalty factor and kernel function on the classification performance of the SVM. The SVM algorithm can be optimized in the process of training; the advantage is that the limit parameters must be stipulated in advance, instead of in the process of model forecast adjustments, according to the classification of the real-time condition, as well as in the process of iteration for super parameters in the process of the training class method, box constraint level, kernel function, the nuclear scale, and standardized data optimization adjustment, until the best classification effect is obtained.

3. Index Selection and Variance Analysis of Rock Burst Intensity Prediction

3.1. Prediction Index Selection

The rock burst mechanism is complex and affected by many factors. This study is based on the characteristics of rock burst, the causes and the internal and external conditions, combining previous researchers’ experience, after a comprehensive analysis to determine the rock burst tendency prediction evaluation index as follows: (1) the maximum tangential stress of surrounding rock and rock uniaxial compressive strength ratio characterizes the surrounding rock stress and rock burst cavern; (2) the ratio of uniaxial compressive strength to the uniaxial tensile strength of rock represents the relationship between rock burst and lithology; (3) the elastic energy index represents the energy characteristics of the rock. Rock burst intensity can be divided into four categories: N (no rock burst activity), L (mild rock burst activity), M (moderate rock burst activity), and H (severe rock burst activity). Referring to the research results of Wang Yuanhan [7], the corresponding relationship between each evaluation index and rock burst intensity is shown in Table 1.

3.2. Establish a Learning Sample Set

By referring to a large amount of literature and relevant materials [5,8], 132 typical rock burst samples were collected and sorted in this paper, and 18 unqualified examples were excluded (e.g., the intensity levels of models in different kinds of literature were inconsistent, etc.). The remaining 114 learning examples are from several projects at home and abroad; they include: Tianshengqiao secondary hydropower station, Longyangxia hydropower station, Lijiaxia hydropower station, Norway Heavy Metal Tunnel, Rail Lead-zinc sulfide mine in Italy, Qinling tunnel, Jiangbian Hydropower Station, Jinchuan No. 2 mine, Maluping mine, Beizhihe Iron Mine, Jinping Secondary power station, Cangling tunnel, Erlang Mountain tunnel, and so on. Specific data parameters of learning samples for rock burst intensity classification prediction are shown in Table 2.

3.3. Classification Index Analysis

Theoretically, the three rock burst intensity classification prediction indexes selected in this paper reflect the characteristic information of rock burst from different angles. Generally, there is no correlation between indicators. To verify the reliability of indicator selection, analysis of variance between indicators and Spearman’s correlation coefficient hypothesis testing analysis were carried out to show the relationship between the three groups of hands.

3.3.1. Analysis of Indicator Variance

Figure 1 and Table 3 shows the results of the analysis of indicator variance based on the data of 114 groups of learning samples, where p = 1.7090 × 10−90, which is far less than 0.01. The null hypothesis is rejected and highly significant differences between the three indicators are considered, indicating that the selected indicators and learning samples have good representativeness and accuracy.
Table 3 shows the ANOVA Table data presentation of the test data. Figure 1 shows the Variance relation diagram of Hierarchical Prediction Indicators of the test data under the three prediction indicators. In Figure 1, the central red line represents the median of the data for this predictor, the blue box represents the data for half of the central distribution, and the ‘+’ represents the outliers.

3.3.2. Spearman’s Correlation Coefficient Hypothesis Test

Hypothesis testing using Spearman’s correlation coefficient should meet the following conditions: (1) The data to be tested are usually a population conforming to normal distribution; (2) The absolute Euclidean distance between test data should not differ too much; (3) Each group of samples was sampled independently.
Table 4 shows the mathematical symbolic representation of the variables represented by the three hierarchical indexes. For the convenience of data processing, it is declared that the hierarchical indexes 1, 2 and 3 will be used for data processing and simulation. Table 5 uses statistical principles to perform basic statistical calculations on the three hierarchical indexes and visualize the data.
Figure 2 shows the Q–Q (Quantile) graph of classification index data. The Q–Q (Quantile) graph in statistics is used to compare the quantiles of the sample data’s probability distribution. The sample data identified are approximately a straight line, so it is concluded that the sample data have an approximately normal distribution, which meets the first requirement.
It can be seen from Figure 3 and Figure 4 that the classification index data are independent and have a relatively stable variation trend, and the error is within an acceptable range, meeting the requirements of Articles 2 and 3. Therefore, Spearman’s correlation coefficient can be used for hypothesis testing.
Table 6 shows the Analysis Results of Spearman’s Index for grading Indicators. The null hypothesis H 0 :   r = 0   and   the   alternative   hypothesis   H 1 : r 0 were established to test whether the correlation coefficient was significantly different from 0. The analysis results are shown in Table 4, where rejecting the null hypothesis means that the correlation coefficient is quite different from 0. As a result, classification index 1 and the grading index correlation are substantially different from those of classification index 3 and classification index 2, respectively, with hierarchical index grade indices for 1 and 3 not being significantly different from zero. Still, the correlation coefficient is below 0.01, therefore the differences between the three classification indexes show that selection of indicators and learning samples have an excellent representative and accuracy.
In summary, the analysis of the above two methods shows that there is a high degree of difference between the three indicators based on the index analysis results of 114 groups of learning sample data, indicating that the selected indicators and learning samples have good representativeness and accuracy. The SVM hierarchical prediction discriminant model has good accuracy and persuasiveness.

4. Optimizable Support Vector Machine Model for Rock burst Intensity

4.1. Optimized Support Vector Machine Discriminant Analysis

The main idea of the SVM discriminant analysis method is based on the existing rock burst sample data, mining the internal relationship between the evaluation index and rock burst intensity, and establishing a discriminant model for judging new samples. Due to the mature application of this method in other engineering predictions, relevant contents can be referred to in [9,10,11].

4.2. Establishment and Verification of the Discriminant Analysis Model of the Optimized Support Vector Machine

According to the SVM discriminant analysis model, three rock burst intensity prediction indexes were taken as input units, 114 groups of rock burst samples were formed into learning sample sets, and four kinds of rock burst grades were taken as different output units. The SVM discriminant analysis method for rock burst intensity classification prediction proposed in this paper was trained, verified and predicted. Through the realization of the algorithm, various parameters of the optimized SVM discriminant analysis model are obtained and the results are shown in Table 7.
Figure 5, Figure 6 and Figure 7 show the data distribution among the three hierarchical indexes of each sample point, which can show the detailed data of the misclassified points.
Figure 8 is the Minimum Classification Error diagram, which shows how the Minimum classification error value changes during the training process. Figure 9 shows the Confusion matrix diagram of Validation data, which shows the classification efficiency and success rate of the paper’s method.
As seen from the above-illustrated results, the discriminant results of samples numbered 7, 16, 100, 107 and 110 are inconsistent with the actual situation. The first three are misjudged as type III rock bursts (in the actual results, type III rock bursts account for 40% of the total number), and the overall misjudgment rate is only 4.38%. The discriminant error can be eliminated gradually by increasing the number of learning samples and strengthening model training. In addition, the cross-validation method is considered in this paper for model testing. Test results are used as performance indicators for the classifier or model. The goal is to achieve high predictive accuracy and low predictive error. To ensure the accuracy and stability of the cross-validation results, it is necessary to divide a test data set into different complementary subsets multiple times and cross-validate repeatedly. Take the average value of numerous validations as the validation result. The cross-validation method makes each sample data to be used as training and test data, which can effectively avoid the occurrence of over-learning and under-learning states. Through actual measurement, it is found that the accuracy of the validation data obtained by using the cross-validation method is 95.6% higher than that of the model tested by using the set-aside method, which is 92.9%. Among them, the proportion of the training set and the test set of the set-aside method is 75%:25%, and the results of the reserved method change slightly after the model is run, so it is difficult to avoid the instability of the single use of the mysterious method, which not only wastes data, but is also susceptible to overfitting, and the model correction method is not convenient. Therefore, the results obtained using the cross-validation method for model verification are accurate and persuasive.

5. Application of Engineering Examples

The trained distance discriminant model is used to classify and predict the rock burst intensity of 10 projects, respectively. The three prediction index data of samples are from the 1–1 section of the diversion tunnel of Jinping Ii Power Station [11], level 255 m of the Tongkang Mine of the China Tin Group [12], Jinchuan Ii Ore circle, Daxiangling Tunnel YK61 + 305 [13], Dongguashan Copper Mine [14] and other domestic and foreign projects [15]. Table 6 shows that the classification prediction results of the SVM discriminant model of rock burst intensity established in this paper are entirely in agreement the actual situation and have significant accuracy with the prediction results of other prediction models [15].
Table 8 shows Rock Burst Instance data and Classification Prediction results. The accuracy of Optimization of SVM is 100% for ten samples randomly selected from different countries and different geology, which is obviously better than other representative methods. Figure 10 shows the Confusion matrix diagram of actual test data, Figure 11 shows TPR and FNR Predicted Results of Actual test data, The two graphs together show that all ten random sample points were successfully and accurately predicted.
Compared with other mathematical model methods, the SVM discriminant model does not need to consider various human factors, such as index membership degree and weight coefficient, it has high learning efficiency, and it is easy to use. It only needs to substitute the data of the samples to be predicted into the discriminant model. It can be seen that the SVM discriminant model for rock burst intensity classification prediction established in this paper based on extensive sample data has high accuracy and practicability, which can well meet the needs of engineering and can be popularized and applied.

6. Conclusions

(1)
A total of 114 groups of rock burst samples are selected, and the analysis of variance and Spearman’s correlation coefficient hypothesis test shows that the correlation between the selected three indicators is weak and has good representativeness.
(2)
The algorithm model is established based on rock burst in all kinds of data at home and abroad; therefore, the sample data are suitable for all kinds of geological conditions of rock burst prediction. In practical engineering applications, the random selection is different from the original sample data, in the case of forecasting the different geological conditions of rock burst prediction data at home and abroad. The accuracy of the final prediction results and the actual results reaches 100.00%, which shows that the algorithm is suitable for rock burst prediction under various geological conditions and has a certain adaptability.
(3)
The judgment accuracy of the optimized SVM discriminating model established was 95.6%. In the future, it is necessary to extensively collect many representative rock burst case data, expand and perfect the training sample database, and continuously improve the classification prediction accuracy.
(4)
The optimizable support vector machine (SVM) model, as a supervised learning algorithm, greatly eliminates the influence of the choice of penalty factor and kernel function on the classification performance of the SVM. The SVM algorithm can be optimized in the process of training; the advantage is that the limit parameters must be stipulated in advance, instead of in the process of model forecast adjustments, according to the classification of the real-time condition, as well as in the process of iteration for super parameters in the process of training class method, box constraint level, kernel function, the nuclear scale, and standardized data optimization adjustment, until the best classification effect is obtained.
(5)
The optimizable support vector machine model selected in this paper breaks through the experience and randomness of setting the hyperparameters of the model, and the hyperparameter optimization based on Bayesian optimization makes the accuracy of the support vector machine model reach 95.6% and 100.00% in both the learning and the actual parts. Future research can be improved by further integrating the actual situation. According to different geological conditions, the establishment and prediction of data sets can be limited to one geological situation, and then combined with the optimized support vector machine model selected in this paper for prediction, and more accurate prediction results can be obtained.

Author Contributions

Conceptualization, S.Y. and Y.Z.; methodology, S.Y. and X.L.; software, S.Y. and R.L.; validation, S.Y., Y.Z. and X.L.; investigation, S.Y. and Y.Z.; resources, Y.Z.; data curation, S.Y. and R.L.; writing—original draft preparation, S.Y., X.L. and R.L.; writing—review and editing, S.Y. and R.L.; visualization, S.Y. and R.L.; supervision, Y.Z.; project administration, S.Y. and Y.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China (Grant No.52074123).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

We gratefully acknowledge the constructive comments from the editor and the anonymous referees. This research was funded by National Natural Science Foundation of China (Grant No.52074123).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Feng, X.-T. Research on Rockburst Breeding Mechanism, Early Warning and Dynamic Control Method; Wuhan Institute of Rock and Soil Mechanics, Chinese Academy of Sciences: Wuhan, China, 2012. [Google Scholar]
  2. Feng, X.-T. Rockburst: Mechanisms, Monitoring, Warning, and Mitigation; Elsevier: Amsterdam, The Netherlands, 2017. [Google Scholar]
  3. Chinanews. Yunnan Channel. A Summary of Recent Tunnel Collapses [EB/OL]. Available online: http://www.yn.chinanew.cial/2014/0916/20967.html (accessed on 16 September 2014).
  4. Li, P.; Chen, B.; Zhou, Y. Research progress of rockburst prediction and early warning in hard rock underground engineering. J. China Coal Soc. 2019, 44, 447–465. [Google Scholar]
  5. Liu, X.; Ji, H. Research on Rockburst Prediction Based on AdaBoost-BAS-SVM Model. Metal Mine 2021, 10, 28–34. [Google Scholar] [CrossRef]
  6. Shahriari, B.; Swersky, K.; Wang, Z.; Adams, R.P.; de Freitas, N. Taking the Human Out of the Loop: A Review of Bayesian Optimization. Proc. IEEE 2016, 104, 148–175. [Google Scholar] [CrossRef]
  7. Wang, Y.; Li, W.; Lee, P.K.K.; Tsui, Y.; Tham, L.G. Method of fuzzy comprehensive evaluations for rockburst prediction. Chin. J. Rock Mech. Eng. 1998, 5, 15–23. Available online: https://kns.cnki.net/kcms/detail/detail.aspx?FileName=YSLX805.002&DbName=CJFQ1998 (accessed on 1 August 2022).
  8. Wang, C.; Li, Y.; Zhang, C. Prediction Model of Rockburst Intensity Classification Based on Data Mining Analysis with Large Samples. J. Kunming Univ. Sci. Technol. Nat. Sci. 2020, 45, 26–31. [Google Scholar] [CrossRef]
  9. Wang, H.; Lu, Z.; Qiu, J.; Liu, W.; Zhang, L. Prediction of Rock Burst by Improved Particle Swam Optimization based Support Vector Machine. Chin. J. Undergr. Space Eng. 2017, 13, 364–369. [Google Scholar]
  10. Xu, R.; Hou, K.; Wang, X.; Liu, X.-Q.; Li, X.-B. Combined Prediction Model of Rockburst Intensity Based on Kemel Principal Component Analysis and SVM. Gold Sci. Technol. 2020, 28, 575–584. [Google Scholar]
  11. Zhou, J.; Li, X.B.; Shi, X.Z. Long-term prediction model of rockburst in underground openings using heuristic algorithms and support vector machines. Saf. Sci. 2012, 50, 629–644. [Google Scholar] [CrossRef]
  12. Zhou, K.; Lei, T.; Hu, J. RS-Topsis model of rockburst prediction in deep metal mines and its application. Chin. J. Rock Mech. Eng. 2013, 32, 3705–3711. [Google Scholar]
  13. Yi, Y.; Cao, P.; Pu, C. Multi-factorial Comprehensive Estimation for Jinchuan’s Deep Typical Rockburst Tendency. Sci. Technol. Rev. 2010, 28, 76–80. [Google Scholar]
  14. Zhang, J. Study on Prediction by Stages and Control Technology of Rock Bursthazard of Daxiangling Highway Tunnel. Ph.D. Thesis, Southwest Jiaotong University, Chengdu, China, 2010. Available online: https://kns.cnki.net/kcms/detail/detail.aspx?FileName=2010122032.nh&DbName=CMFD2010 (accessed on 1 August 2022).
  15. Xie, X.; Li, D.; Kong, L.; Ye, Y.; Gao, S. Rockburst propensity prediction model based on CRITIC-XGB algorithm. Chin. J. Rock Mech. Eng. 2020, 39, 1975–1982. [Google Scholar] [CrossRef]
Figure 1. Variance relation diagram of hierarchical prediction indicators.
Figure 1. Variance relation diagram of hierarchical prediction indicators.
Mathematics 10 03276 g001
Figure 2. Q–Q (Quantile) graph of classification index data.
Figure 2. Q–Q (Quantile) graph of classification index data.
Mathematics 10 03276 g002
Figure 3. Fractal dimension visualization of grading indicators.
Figure 3. Fractal dimension visualization of grading indicators.
Mathematics 10 03276 g003
Figure 4. The relationship and trend of changes among hierarchical prediction indexes.
Figure 4. The relationship and trend of changes among hierarchical prediction indexes.
Mathematics 10 03276 g004
Figure 5. Scatter diagram of prediction results in low dimension (1).
Figure 5. Scatter diagram of prediction results in low dimension (1).
Mathematics 10 03276 g005
Figure 6. Scatter diagram of prediction results in low dimension (2).
Figure 6. Scatter diagram of prediction results in low dimension (2).
Mathematics 10 03276 g006
Figure 7. Scatter diagram of prediction results in low dimension (3).
Figure 7. Scatter diagram of prediction results in low dimension (3).
Mathematics 10 03276 g007
Figure 8. Minimum classification error diagram.
Figure 8. Minimum classification error diagram.
Mathematics 10 03276 g008
Figure 9. Confusion matrix diagram of validation data.
Figure 9. Confusion matrix diagram of validation data.
Mathematics 10 03276 g009
Figure 10. Confusion matrix diagram of actual test data.
Figure 10. Confusion matrix diagram of actual test data.
Mathematics 10 03276 g010
Figure 11. TPR and FNR predicted results of actual test data.
Figure 11. TPR and FNR predicted results of actual test data.
Mathematics 10 03276 g011
Table 1. Prediction standard of rock burst tendency.
Table 1. Prediction standard of rock burst tendency.
IndexValues at Each Rock Burst Level
NLMH
σ θ σ c 0~0.30.3~0.50.5~0.7>0.7
σ c σ t >4026.7~4014.5~26.70~14.5
W e t 0~22~44~6>6
Table 2. Rock burst sample data and classification prediction results.
Table 2. Rock burst sample data and classification prediction results.
Sample NumberClassification IndexA LP LSample
Number
Classification IndexA LP L
σ θ σ c σ c σ t W e t σ θ σ c σ c σ t W e t
10.34246.6MM580.5725.63.8MM
20.1131.27.4NN590.6125.63.7MM
30.1235.7NN600.5629.24.8MM
40.4221.75MM610.7132.25.5HH
50.7717.55.5HH620.4949.54.7MM
60.414.77.1MM630.4645.55.2MM
7 *0.5414.26.2HM640.47555MM
80.4157.1MM650.3142.81.8NN
90.5813.26.3HH660.61253.7MM
100.2362.3NN670.5531.34.6MM
110.1947.91.9NN680.6932.15.9HH
120.6613.26.8HH690.550.95.2MM
130.415.63.5LL700.6916.93.4MM
140.4413.12.1LL710.421710.9HH
150.136.71.4NN720.320.45LL
16 *0.37245.1LM730.5412.24.9MM
170.4511.22LL740.211.23.6NN
180.196.71.4NN750.3522.73.3LL
190.48245.1MM760.7213.99.1HH
200.7424.46.3HH770.6414.47.7HH
210.236.71.4NN780.4716.55.5MM
220.61245.1MM790.5218.64.2MM
23111.22HH800.4514.83.1LL
240.289.71.9NN810.214.13.6NN
250.711.72.8MM820.5511.14MM
260.1127.27NN830.5616.33.3MM
270.1318.83.6NN840.4130.74.3LL
280.121.44.7NN850.2294.9LL
290.6726.80.9LL860.456.82.2LL
300.8328.93.2MM870.289.56.1MM
310.9328.93.2HH880.3512.12.9LL
320.7428.93.2MM890.6622.33.2MM
331.4119.23.1HH900.7213.25.2HH
340.79222MM910.3729.73.5LL
350.5620.42LL920.4232.83LL
360.4620.42LL930.2842.72.2NN
370.4919.72.3LL940.3828.83LL
380.4419.72.3LL950.7227.54.3MM
390.8419.72.3MM960.6916.65.7HH
400.4219.72.3LL970.4215.53.2LL
410.4619.72.3LL980.2236.41.8NN
420.2823.64.9LL990.6219.44.5MM
430.2823.84.8LL100 *0.5731.23.2LM
440.5221.15.5MM1010.6512.45.4HH
450.6528.66.8HH1020.5918.84.2MM
460.1129.42NN1030.7329.73.8MM
470.237.51.5NN1040.3742.32.8NN
480.4420.38.1HH1050.4711.113.97MM
490.628.31.8LL1060.5314.115.76MM
500.6417.57.2HH107 *0.489.767.27MH
510.4345.91.7NN1080.4313.987.27MM
520.4229.92.4LL1090.414.737.08MM
530.5634.31.9LL110 *0.5514.726.43MH
540.628.33.4MM1110.61253.7MM
550.53213.6MM1120.5531.34.6MM
560.6621.54.1MM1130.6932.15.9HH
570.5217.84.3MM1140.550.95.2MM
Note: Explanation of Abbreviation Meaning: A L: Actual level; P L: Predicting level. The part that is boldfaced and italicized and marked with * indicates that the predicted results are not consistent with the actual results.
Table 3. ANOVA visualization table of the test data.
Table 3. ANOVA visualization table of the test data.
ANOVA Table
SourcesSSdfMSFp
line31,791.2215,895.6404.311.7090 × 10−90
error13,327.833939.3
total45,119.1341
Table 4. The grading index corresponds to the symbol description.
Table 4. The grading index corresponds to the symbol description.
Grading Index 1 σ θ σ c
Grading Index 2 σ c σ t
Grading Index 3 W e t
Table 5. Statistical parameters of hierarchical index fractal dimension data.
Table 5. Statistical parameters of hierarchical index fractal dimension data.
Descriptive StatisticsGrading Index 1Grading Index 2Grading Index 3
Min0.16.70.9
Max1.415510.9
Mean value0.49184210522.550087724.214736842
Median0.48520.43.985
Coefficient of skewness0.618129430.9244687670.605471435
Kurtosis coefficient5.2729207843.5980729183.192751812
Coefficient of standard deviation0.20829083610.689349381.907852525
Table 6. Analysis results of Spearman’s index for grading indicators.
Table 6. Analysis results of Spearman’s index for grading indicators.
Grading Index 1Grading Index 2Grading Index 3
Spearman’s RhoGrading Index 1correlation coefficient10.039 **0.135
Sig.00.6780.151
Number114114114
Grading Index 2correlation coefficient0.039 **1−0.033 ***
Sig.0.67800.729
Number114114114
Grading Index 3correlation coefficient0.135−0.033 ***1
Sig.0.1510.7290
Number114114114
***, p < 0.01 means rejecting the null hypothesis at the 99% confidence level. **, p < 0.05 means rejecting the null hypothesis at the 95% confidence level.
Table 7. Training results of optimized SVM model parameters.
Table 7. Training results of optimized SVM model parameters.
Training ResultsAccuracy (Validation)95.6%
Total cost (Validation)5
Prediction speed~6800 obs/s
Training time39.308 s
Test ResultsAccuracy (Test)100.0%
Total cost (Test)0
Model TypePresetBayesian optimization
Table 8. Rock burst instance data and classification prediction results.
Table 8. Rock burst instance data and classification prediction results.
Sample NumberClassification IndexA LOptimization of SVM
P L
XGBoot Model P LRF Model P LSVM Model P L
σ θ σ c σ c σ t W e t
10.62203.1MMH *MM
20.6117.95.3MMMMM
30.4413.12.1LLM *M *L
40.7132.25.5HHHHH
50.47114MMH *MH *
60.2520.773.80NNNL *N
70.1520.773.80NNNNN
80.5813.206.30HHHHH
90.4517.505.10MMMH *H *
100.6613.206.80HHHHH
Precision rate100%70%70%80%
Note: A L: Actual level; P L: Predicting level. The intensity level marked * does not match the actual level.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Yan, S.; Zhang, Y.; Liu, X.; Liu, R. Rock Burst Intensity Classification Prediction Model Based on a Bayesian Hyperparameter Optimization Support Vector Machine. Mathematics 2022, 10, 3276. https://doi.org/10.3390/math10183276

AMA Style

Yan S, Zhang Y, Liu X, Liu R. Rock Burst Intensity Classification Prediction Model Based on a Bayesian Hyperparameter Optimization Support Vector Machine. Mathematics. 2022; 10(18):3276. https://doi.org/10.3390/math10183276

Chicago/Turabian Style

Yan, Shaohong, Yanbo Zhang, Xiangxin Liu, and Runze Liu. 2022. "Rock Burst Intensity Classification Prediction Model Based on a Bayesian Hyperparameter Optimization Support Vector Machine" Mathematics 10, no. 18: 3276. https://doi.org/10.3390/math10183276

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop