Slope Stability Prediction Method Based on Intelligent Optimization and Machine Learning Algorithms

: Slope engineering is a type of complex system engineering that is mostly involved in water conservancy and civil and mining engineering. Moreover, the link between slope stability and engineering safety is quite close. This study took the stable state of the slope as the prediction object and used the unit weight, cohesion, internal friction angle, pore water pressure coefﬁcient, slope angle, and slope height as prediction indices to analyze the slope stability based on the collection of 117 slope data points. The genetic algorithm was used to solve the hyperparameters of machine learning algorithms by simulating the phenomena of reproduction, hybridization, and mutation in the natural selection and natural genetic processes. Five algorithms were used, including the support vector machine, random forest, nearest neighbor, decision tree, and gradient boosting machine models. Finally, all of the obtained stability prediction results were compared. The prediction outcomes were analyzed using the confusion matrix, receiver characteristic operator (ROC), and area under the curve (AUC) value. The AUC values of all machine learning prediction results were between 0.824 and 0.964, showing excellent performance. Considering the AUC value, accuracy, and other factors, the random forest algorithm with KS cutoff was determined to be the optimal model, and the relative importance of the inﬂuencing variables was studied. The results show that cohesion was the factor that most affects slope stability, and the inﬂuence factor was 0.327. This study proves the effectiveness of the integrated techniques for slope stability prediction, makes essential suggestions for future slope stability analysis, and may be extensively applied in other industrial projects.


Introduction
Slope stability analysis is among the most critical aspects of geotechnical engineering [1] as it impacts the safety of industrial engineering projects [2]. Slope stability analysis is a more complex and comprehensive work, which is riskier and more challenging than other geotechnical engineering projects. Slope instability is a complex natural phenomenon that causes severe natural disasters and economic losses in many countries [3]. Therefore, building a safe, reliable, and effective model for slope stability analysis, evaluation, and prediction is crucial for reducing the geological risk of slope disasters and assuring the safety of people and property.
There are numerous ways to analyze a slope's stability, the earliest of which primarily uses qualitative analysis techniques [4]. With the development of computer technology and the in-depth study of rock mechanics mechanisms [1], quantitative analysis methods are gradually applied. However, it has not reached a degree of complete quantification for slope stability analysis but is only limited to semi-quantitative analyses. The main methods are the limit equilibrium method and numerical simulation method. The limit equilibrium method determines the stability by analyzing the safety factor under the critical failure state of the slope. The widely used limit equilibrium methods include the Bishop method [5], Janbu method [6], Morgenstern-Price method [7], Sarma method [8], etc. So far, the limit equilibrium method has been frequently employed in calculating slope stability [9][10][11]. Despite its ease of use, it does not accurately reflect actual soil stress; hence, the sliding surface and blocks need to be assumed. Accordingly, the finite element numerical simulation method based on elastic-plastic theory has been widely used in slope stability analyses [12][13][14][15]. However, the accuracy of the finite element numerical simulation results mainly depends on the accuracy of the input parameters and selected calculation model. Its parameters and model are difficult to determine, so this method also has limitations [16]. In recent years, machine learning-based techniques have been applied to slope stability analyses due to scientific and technological advancements. Such methods analyze the relationship between slope stability and its influence by analyzing and predicting slope stability using the available slope data. Zhang and Luo [16] proposed a model for predicting the stability of slopes using an artificial immune algorithm based on the antigen recognition mechanism of the biological immune system. The backpropagation neural network, naive Bayes, decision tree (DT), and support vector machine (SVM) models are used to analyze the slope stability of open-pit mines [17]. Researchers [18,19] predicted the minimum safety factor of different soil slopes through an artificial neural network (ANN). Sun et al. [20] searched for the critical slip surface of the slope through a genetic algorithm and spline curve and calculated the safety factor of the slope. Based on the opposition-based learning method, Khajehzadeh et al. [21] improved the firefly algorithm (FA) and applied it to the slope stability analysis application to find the critical failure surface.
However, the machine learning algorithm needs to set hyperparameters, and an appropriate hyperparameter setting will make the results more accurate. Therefore, different scholars have carried out hyperparameters optimization according to the situation. Bui et al. [22] and Zhang et al. [23] used particle swarm optimization (PSO) to optimize the hyperparameters of k-nearest neighbor (KNN) and extreme gradient boosting algorithms, respectively, both of which are used to predict the ground vibration induced by explosions. Khajehzadeh et al. [24] used the improved sine cosine algorithm to optimize the ANN model to predict the slope stability coefficient under static and dynamic load conditions. Gordan et al. [25] used the ANN and PSO models to predict the slope stability coefficient during earthquakes. Cheng and Hoang [26] used the artificial bee colony algorithm to optimize support vector classification to analyze the slope stability analysis caused by typhoons. Cheng and Hoang [27] and Hoang and Pham [28] used the FA model to optimize the least squares support vector classification and the KNN models, respectively, to establish a comprehensive slope prediction model for slope stability analysis. Gao et al. [29] optimized the ANN hyperparameters through the imperialist competition algorithm to predict the slope stability of cohesive soil.
Although the heuristic intelligent optimization algorithm performs well for optimizing hyperparameters, this algorithm has a poor performance when dealing with discrete problems. It is easy to fall into local optimization, and thus the algorithm cannot find the optimal hyperparameters well. This paper selects the genetic algorithm based on natural evolution theory for hyperparametric optimization. This algorithm is used for slope stability analysis, and there is little research on hyperparametric optimization using genetic algorithms. This algorithm can effectively avoid the problem that other heuristic algorithms easily fall into regarding local convergence, and it can better find the global optimal solution [30]. It has high parallelism and good robustness, and it directly takes the objective function value, i.e., the hyperparametric value, as the search information. The genetic algorithm is more suitable for solving complex optimization problems such as hyperparametric optimization, and the model has a good effect on hyperparametric optimization in other industrial fields [31][32][33]. This research effectively combines the genetic algorithm with slope stability prediction, which can reflect the rationality of artificial intelligence (AI) technology for solving engineering problems and is of great significance for slope stability design and other engineering problems.

Machine Learning and Genetic Algorithm
In this study, the machine learning algorithm was used to simulate and predict slope stability, and an evolutionary genetic algorithm was used to optimize the hyperparameters of the machine learning algorithm. This chapter briefly introduces the principles of the five machine learning and genetic algorithms used in this paper.

Machine Learning
In this research, five machine learning algorithms re selected. The reason for choosing these five machine learning algorithms, namely the support vector machine (SVM), random forest (RF), k-nearest neighbor (KNN), decision tree (DT), and gradient boosting machine (GBM) algorithms, are the algorithms that can better deal with binary or multi-classification problems that are affected by multiple factors [34]. All need to select reasonable hyperparameters so that the algorithm itself has better prediction results and the algorithm has a better application in the industrial field.

Support Vector Machine
The SVM model is a binary classification model that finds a hyperplane to segment the samples. The basic idea of segmentation is to maximize the interval to as large as possible while simultaneously transforming it into a convex quadratic programming problem. The core is to maximize the classification boundary, as shown in Figure 1. hyperparametric optimization, and the model has a good effect on hyperparametric opti mization in other industrial fields [31][32][33]. This research effectively combines the geneti algorithm with slope stability prediction, which can reflect the rationality of artificial in telligence (AI) technology for solving engineering problems and is of great significanc for slope stability design and other engineering problems.

Machine Learning and Genetic Algorithm
In this study, the machine learning algorithm was used to simulate and predict slop stability, and an evolutionary genetic algorithm was used to optimize the hyperparame ters of the machine learning algorithm. This chapter briefly introduces the principles o the five machine learning and genetic algorithms used in this paper.

Machine Learning
In this research, five machine learning algorithms re selected. The reason for choosing these five machine learning algorithms, namely the support vector machine (SVM), ran dom forest (RF), k-nearest neighbor (KNN), decision tree (DT), and gradient boosting ma chine (GBM) algorithms, are the algorithms that can better deal with binary or multi-clas sification problems that are affected by multiple factors [34]. All need to select reasonabl hyperparameters so that the algorithm itself has better prediction results and the algo rithm has a better application in the industrial field.

Support Vector Machine
The SVM model is a binary classification model that finds a hyperplane to segmen the samples. The basic idea of segmentation is to maximize the interval to as large as pos sible while simultaneously transforming it into a convex quadratic programming prob lem. The core is to maximize the classification boundary, as shown in Figure 1. The interval is equal to the projection of the difference between two heterogeneou support vectors on W, i.e., where → and → represent two positive and negative support vectors, respectively, and where → and → satisfies ( + ) = 1, i.e., The interval is equal to the projection of the difference between two heterogeneous support vectors on W, i.e., x − represent two positive and negative support vectors, respectively, and where → x + and → x − satisfies y i (w T x i + b) = 1, i.e., Thus, the interval is determined, and the core idea of SVM is to maximize the interval, that is: max Maximizing 2 ||w|| is equivalent to minimizing || w ||, i.e., The above formula is the basic model of an SVM approach. As SVM obtains much better results than other algorithms in limited training set scenarios, it is extensively applied for slope stability prediction and has a significant prediction effect [35,36].

Decision Tree
The DT model is a fundamental classification algorithm of machine learning; it predicts the value of a target variable by learning simple decision rules from data attributes. The implementation process of the algorithm includes feature selection, DT generation, and DT pruning. Firstly, the algorithm classifies according to the characteristics and then recursively generates the DT. At the same time, it prunes the generated DT to cutoff the redundant and useless branches. The DT generation corresponds to the model's local optimal solution, and the pruning of the DT corresponds to the global optimal solution. The DT is intuitively understood and applied in the actual slope prediction. Hwang et al. [37] used the DT algorithm to analyze and classify the factors affecting the slope. The results show that the DT has a good classification effect.

Random Forest
The essence of the RF model is that it is an enhancement to the DT algorithm [38]. It functions by creating several decision trees. Using the bootstrap resampling technique, m random samples are continuously extracted from the original training sample set N. On the basis of the self-help sample set, a new training sample set is generated, and m classification trees are constructed to form an RF. The new data classification is based on the score derived from the number of votes in the classification tree. The model's schematic diagram is shown in Figure 2. The reason for choosing the RF model was that it has been proven to be one of the most robust algorithms and has been proven to have high accuracy in many data sets [39]; the RF model is also well-used in classification problems [40].
Thus, the interval is determined, and the core idea of SVM is to maximize the interval that is: Maximizing || || is equivalent to minimizing || w ||, i.e., The above formula is the basic model of an SVM approach. As SVM obtains much better results than other algorithms in limited training set sce narios, it is extensively applied for slope stability prediction and has a significant predic tion effect [35,36].

Decision Tree
The DT model is a fundamental classification algorithm of machine learning; it pre dicts the value of a target variable by learning simple decision rules from data attributes The implementation process of the algorithm includes feature selection, DT generation and DT pruning. Firstly, the algorithm classifies according to the characteristics and then recursively generates the DT. At the same time, it prunes the generated DT to cutoff th redundant and useless branches. The DT generation corresponds to the model's local op timal solution, and the pruning of the DT corresponds to the global optimal solution. Th DT is intuitively understood and applied in the actual slope prediction. Hwang et al. [37 used the DT algorithm to analyze and classify the factors affecting the slope. The result show that the DT has a good classification effect.

Random Forest
The essence of the RF model is that it is an enhancement to the DT algorithm [38]. I functions by creating several decision trees. Using the bootstrap resampling technique, m random samples are continuously extracted from the original training sample set N. On the basis of the self-help sample set, a new training sample set is generated, and m classi fication trees are constructed to form an RF. The new data classification is based on th score derived from the number of votes in the classification tree. The model's schemati diagram is shown in Figure 2. The reason for choosing the RF model was that it has been proven to be one of the most robust algorithms and has been proven to have high accuracy in many data sets [39]; the RF model is also well-used in classification problems [40].

Nearest Neighbor Classification Algorithm
Among data mining classification techniques, the KNN (k-nearest neighbor) classification algorithm is one of the simplest approaches. The KNN nearest neighbor classification algorithm is implemented by taking all samples of known categories as references, calculating the distance between known and unknown samples, selecting the known samples (K) closest to the unknown samples, classifying the unknown samples and nearest samples (k) into one category. The KNN algorithm is simple and easy to understand and implement. The algorithm is not sensitive to outliers, which makes the prediction effect suitable and applicable for use [41].

Gradient Boosting Machine
The GBM model is quite common type of integrated learning algorithm; it generates multiple trees through multiple features to make decisions. Each tree learns the conclusions and residuals of all trees before learning, and it reduces the loss by fitting the last residuals. The gradient tree boosting is a generalization of the lifting algorithm for any differentiable loss function and is applied to classification and regression applications. The results of slope stability predictions using the GBM model are promising [42].

Genetic Algorithm
For the five machine learning algorithms, the setting of hyperparameters significantly impacts the results. In addition to setting parameters based on experience, certain algorithms are often used for optimization processes. However, this has the drawback of resulting in a local optimal solution. Thus, this study set up an intelligent optimization algorithm to adjust the hyperparameters to solve said problem. An intelligent optimization algorithm, also known as one that is meta-heuristic, is generally a random search algorithm based on biological intelligence or physical phenomena; examples of these are the simulated annealing algorithm [43], particle swarm optimization algorithm [44], firefly algorithm [45], ant colony algorithm [46], etc. This kind of new algorithm generally does not require the continuity of objective function and constraint. Therefore, it can be effectively used to adjust the hyperparameters.
The genetic algorithm (GA), a type of evolutionary algorithm, was proposed in 1969 by Professor Holland [47]; it is a heuristic search algorithm based on Darwin's evolution theory of natural selection and population genetics [48]. In natural selection and natural genetics, the algorithm simulates the phenomena of reproduction, hybridization, and mutation [34]. When using the genetic algorithm to solve the problem, some initial populations encoded by a certain length are randomly generated. Each individual is evaluated by their fitness function, and those with high fitness values are chosen to participate in the genetic process, while those with low fitness are eliminated. The collection of individuals forms a new generation of the population through genetic operation (replication, crossover, and variation) until the stopping criterion is met. The best-realized individual in the offspring is taken as the execution result of the genetic algorithm. Figure 3 shows the hyperparameter optimization steps of the genetic algorithm. As a combinatorial optimization approach based on the process of biological evolution, GA has the benefit of being easy to integrate with other systems [34]. GA effectively deals with complex multivariable optimization problems. Its unique cross-recombination, selection, and mutation operations prevent the optimization process from falling into the local optimum or missing the global optimum [49,50]. It also has a good application to the hyperparameter optimization of the machine learning algorithm. Zhou et al. [51] studied the prediction of soil liquefaction potentials by optimizing RF hyperparameters through a genetic algorithm. SVM parameters are selected through different algorithms to classify and identify dangerous railway goods, which have good results [52].

Data
The main factors affecting slope stability are stratum and lithology, geological struc ture and in situ stress, rock mass structure, water action, slope geometry, and surface mor phology. When using the machine learning method for prediction research, the rationality of selecting research variables will directly affect the prediction results. Different scholar [42,[53][54][55] have selected the six factors of rock bulk density, cohesion, internal friction an gle, slope angle, slope height, and pore water pressure as the research variables of slop stability prediction. Their investigations have demonstrated a considerable correlation be tween these variables and slope stability. The six factors are suitable for slope stability predictions. Therefore, this study also selected the six factors for prediction research. In this study, the slope sample data set was counted from [56][57][58][59], and the same data wa deleted. Finally, 117 groups of data are obtained, including 61 groups of slope stability samples and 56 groups of slope failure samples. The relevant statistical characteristic val ues of the data are shown in Table 1

Data
The main factors affecting slope stability are stratum and lithology, geological structure and in situ stress, rock mass structure, water action, slope geometry, and surface morphology. When using the machine learning method for prediction research, the rationality of selecting research variables will directly affect the prediction results. Different scholars [42,[53][54][55] have selected the six factors of rock bulk density, cohesion, internal friction angle, slope angle, slope height, and pore water pressure as the research variables of slope stability prediction. Their investigations have demonstrated a considerable correlation between these variables and slope stability. The six factors are suitable for slope stability predictions. Therefore, this study also selected the six factors for prediction research. In this study, the slope sample data set was counted from [56][57][58][59], and the same data was deleted.  For the Pearson correlation coefficient, the value was usually in the range of 0.4-0.6. It was considered that the two factors were moderately correlated. If the value was in the range of 0.0-0.2, it was considered to be a very weak correlation or no correlation. The correlation degree of each factor can be seen in the figure, and the nonlinear relationship between each factor can be seen from the discrete figure. Therefore, the data are standardized and scaled to (0,1) to improve the accuracy of the model.

Data Division
In this study, the slope stability prediction belongs to the supervised classifica problem. For this kind of problem, the model is usually established to learn the eigenv and target value. Then, the predicted eigenvalue is the input, and the predicted valu

Data Division
In this study, the slope stability prediction belongs to the supervised classification problem. For this kind of problem, the model is usually established to learn the eigenvalue and target value. Then, the predicted eigenvalue is the input, and the predicted value is the output. Therefore, the data set must be divided into training and testing sets. The selection of the training set must represent the whole data set. If the proportion of the training set is too small, it cannot be correctly learned. When the proportion of the training set is too large, it is easy to produce overfitting. In this study, the proportion of the training set and test set was 70% and 30%, respectively. Hence, the model has sufficient learning and more accurate results.

Evaluation Index
Evaluating machine learning algorithms is vital for solving practical problems. Only through an accurate evaluation can the algorithm be optimized in the later stage. For example, the accuracy, precision, recall, AUC value, and ROC curve are the evaluation indices of the classification algorithm. This paper uses the confusion matrix, AUC value, and ROC curve to evaluate the results. The confusion matrix is an index that evaluates the results of the model. It is the foundation for the ROC curve drawing and a component of the model evaluation as well. The confusion matrix counts the number of observations classified into wrong and proper categories and then displays the results, as seen in Table 2. The true positive (TP) indicates the number of correct predictions of slope stability. The true negative (TN) shows the number of accurate predictions of slope instability. The false positive (FP) indicates the number of wrong forecasts of slope instability. The false negative (FN) indicates the number of incorrect predictions of slope stability. The aforementioned parameters may be used to define three performance indicators: accuracy (ACC), sensitivity (TPR), and specificity (TNR), as described in Table 3.
The ROC curve takes the TPR as the vertical axis and 1-TNR as the horizontal axis to measure the imbalance in classification. The category imbalance frequently happens in actual data sets, i.e., there are more negative samples than positive ones (or vice versa), and the distribution of positive and negative samples in the training data may also alter over time. In this scenario, the ROC curve can remain unchanged. Generally, 50% of the reference line is selected as the cutoff point for the ROC curve drawing. However, the result may have a relative error. Therefore, this paper selects the KS value in the K-S curve as the cutoff point for the ROC curve drawing and compares it with the reference line.

K-Fold Cross Validation
In the hyperparameter adjustment, the prediction model always has the risk of overfitting or selection deviation [60]. Therefore, the CV method is generally used for model optimization. CV is a method used to evaluate the prediction ability of the model on independent data sets, find the hyperparameter value that makes the generalization performance of the model the best, retrain the model on all training sets, and use the independent test set to make the final evaluation of the model performance (see Figure 5). The K-fold divides the data set into k small blocks, and the verification and test sets form complementary sets. This study set 10 fold, the most commonly used CV method.

Hyperparameter Adjustment
Geatpy, a Python toolkit for genetic algorithms, was used to adjust the hyperparameters as it provides an object-oriented evolutionary algorithm framework. When using it for the hyperparameter optimization, we took the learning results of different machine learning algorithms as the objective function, set the algorithm parameters, called the algorithm operator, and finally gave the hyperparameter adjustment results of different algorithms. The value characteristics and range of optimization parameters for different algorithms are shown in Table 4, in which RF, DT, and GBM take common parameters. The hyperparameters and their adjustment range depended on the characteristics of the slope data set and the suggestions of other studies [61,62]. See Figure 6 below to compare the overall experimental process and previous studies.

Hyperparameter Adjustment
Geatpy, a Python toolkit for genetic algorithms, was used to adjust the hyperparameters as it provides an object-oriented evolutionary algorithm framework. When using it for the hyperparameter optimization, we took the learning results of different machine learning algorithms as the objective function, set the algorithm parameters, called the algorithm operator, and finally gave the hyperparameter adjustment results of different algorithms. The value characteristics and range of optimization parameters for different algorithms are shown in Table 4, in which RF, DT, and GBM take common parameters. The hyperparameters and their adjustment range depended on the characteristics of the slope data set and the suggestions of other studies [61,62]. See Figure 6 below to compare the overall experimental process and previous studies.

Parameter Adjustment of Hyperparameters
As mentioned above, this study used the Geatpy toolbox and the principle of the genetic algorithm to adjust the hyperparameters of the five machine learning algorithms. The parameters of the genetic algorithm are selected as follows: the population size is 50, Figure 6. Comparison between previous studies and the present study.

Parameter Adjustment of Hyperparameters
As mentioned above, this study used the Geatpy toolbox and the principle of the genetic algorithm to adjust the hyperparameters of the five machine learning algorithms. The parameters of the genetic algorithm are selected as follows: the population size is 50, the mutation probability is 0.005, and the recombination probability is 0.7. Because the evolutionary algorithm is based on evolution and heredity, in essence, in the optimization process-and it is a random optimization in the evolution process-a genetic algorithm cannot guarantee the stability of the results of multiple runs. Therefore, to analyze the effectiveness of heredity in the hyperparameter adjustment, when the maximum number of iterations is sufficient to produce the optimal AUC value in the evolution process, the optimal AUC value after eight times of optimization is tracked. In addition, we ensured that there was no overfitting and underfitting in the optimization calculation of each model so that the AUC value gradually converges and becomes optimal with the increase in the number of iterations. The average AUC value was obtained and is shown in Table 5. The color gradation table shows that the optimization results of the genetic algorithm for different machine learning algorithms are relatively stable. The average AUC values of the five algorithms were above 0.85, which has good effects. The different algorithms had various performances in the optimal AUC values. Among them, the optimization effect of the SVM algorithm was the best. The average AUC value was 0.956, and the KNN algorithm greatly fluctuated. However, the overall average AUC value tended to be 0.858 for RFC and DT. The average AUC values of the GBM algorithm were 0.938, 0.899, and 0.860, respectively. The optimal hyperparameters of the different machine learning algorithms optimized by the statistical genetic algorithm are shown in Table 6: Table 6. Results of the hyperparameter optimization.

ML Algorithms
Optimum Value

Prediction of Different Machine Learning Algorithms
This section will discuss the prediction performance of the five machine learning methods for the test set and compare their confusion matrix, ROC curve, and AUC value. Table 7 describes the confusion matrix and calculation accuracy of the five machine learning algorithms on the test set when using a 50% baseline cutoff and a KS value cutoff of the K-S curve. The performance of a machine learning algorithm using a KS value cutoff point is generally better than one using a baseline cutoff. The average ACC of all machine learning algorithms increased by about 5.3% from the baseline to the KS value cutoff. The optimized RF model had the highest accuracy, and the RF accuracy of the KS cutoff was higher than that of the baseline cutoff point, from 89% to 94%. The precision of the optimized SVM model with the KS cutoff improved the most from 72% to 86%, while the precision of the optimized DT and KNN models did not change. The optimal RF model with the KS value cutoff point had the highest accuracy. In this case, 34 of the 36 slope cases were correctly predicted. The TPR and TNR calculated from the above table are shown in Figure 7. The best classification model for TPR was the optimized SVM model with a TPR value of 1; in this case, all slope stability conditions were correctly predicted, and it was followed by the optimized RF model. The TPR value of the RF model and KNN model was independent of the selection of the cutoff point. For the optimized DT and GBM models, the TPR without the KS cutoff point had better performance. In contrast, the best prediction of the TNR was realized by the optimized KNN algorithm. Eighteen slope instability models were correctly predicted, but their TNR values were still independent of the selection of cutoff points. The TNR values of the other four algorithms increased with the selection of cutoff points of KS values. As no algorithm had the highest TPR and TNR, the model should be chosen according to AUC value, accuracy, and real engineering needs. Figure 8 shows the ROC curves and AUC values of the five algorithms on the test set. In the whole chart, no classification model dominated the other classification models. However, the optimized SVM and RF models were closer to the upper left part of the coordinate axis than the other algorithms, indicating that they achieved higher overall performances. The AUC values of the two algorithms were also the best, which were 0.964 and 0.944, respectively. The performance of the DT test set was slightly inferior to RF in terms of AUC values. However, the GBM model was even worse, which shows that the RF model had the highest performance in classification based on tree structure integration algorithms (RF, GBM, and DT). In addition, the AUC value of KNN was 0.887, which also has a good effect. To verify the performance of the proposed model, its performance was compared with the results obtained by the benchmark methods, such as the RF, SVM [54], KNN, GBM [63], and DT [17]  out the KS cutoff point had better performance. In contrast, the best prediction of the TNR was realized by the optimized KNN algorithm. Eighteen slope instability models wer correctly predicted, but their TNR values were still independent of the selection of cutof points. The TNR values of the other four algorithms increased with the selection of cutof points of KS values. As no algorithm had the highest TPR and TNR, the model should b chosen according to AUC value, accuracy, and real engineering needs.  Figure 8 shows the ROC curves and AUC values of the five algorithms on the test se In the whole chart, no classification model dominated the other classification models However, the optimized SVM and RF models were closer to the upper left part of th coordinate axis than the other algorithms, indicating that they achieved higher overal performances. The AUC values of the two algorithms were also the best, which were 0.96 and 0.944, respectively. The performance of the DT test set was slightly inferior to RF in terms of AUC values. However, the GBM model was even worse, which shows that th RF model had the highest performance in classification based on tree structure integratio algorithms (RF, GBM, and DT). In addition, the AUC value of KNN was 0.887, which als has a good effect. To verify the performance of the proposed model, its performance wa compared with the results obtained by the benchmark methods, such as the RF, SVM [54] KNN, GBM [63], and DT [17]   Overall, the best RF model with the KS cutoff produced better results than the o four algorithms, and it achieved the accuracy of AUC values of 0.944 and 0.94 and hig TNR values. Although the performance of the SVM algorithm for AUC value and value was slightly higher than that of RF, due to the poor prediction of TNR by SVM (T = 0.76), it is far from the RF model. Considering the reality of slope engineering proble the wrong estimation of the unstable slope will cause severe loss of life and prope Therefore, in slope stability analysis, the correct estimation of TNR is more critical t TPR. Therefore, this paper recommends the optimal RF model with a K-S cutoff poin Overall, the best RF model with the KS cutoff produced better results than the other four algorithms, and it achieved the accuracy of AUC values of 0.944 and 0.94 and higher TNR values. Although the performance of the SVM algorithm for AUC value and TPR value was slightly higher than that of RF, due to the poor prediction of TNR by SVM (TNR = 0.76), it is far from the RF model. Considering the reality of slope engineering problems, the wrong estimation of the unstable slope will cause severe loss of life and property. Therefore, in slope stability analysis, the correct estimation of TNR is more critical than TPR. Therefore, this paper recommends the optimal RF model with a K-S cutoff point for slope stability prediction.

Relative Importance of Influencing Variables
This section further studies the effect of each influencing variable on slope stability. Therefore, three machine learning algorithms (RF, DT, and GBM) with the best performance in AUC value were selected for analysis. Among them, the SVM algorithm performed well in the AUC value. However, SVM was not selected because it could not obtain a characteristic ranking when the RBF kernel function was selected [64]. The final importance ranking of the influencing factors was determined by averaging the results of the three models, as shown in the figure below.
The results of the importance of the influencing variables in the data set of this study ( Figure 9) show that cohesion was the most sensitive factor affecting slope stability in this study. The influence characteristic coefficient is 0.327, accounting for one-third of the total influencing factors, which is consistent with the results that cohesion greatly affects the slope safety coefficient [65], followed by slope height (0.175), rock bulk density (0.16), slope angle (0.142), internal friction angle (0.128), and pore water pressure (0.068). Pore water pressure had the smallest impact on the slope, with a value of only 0.068. This conclusion is basically the same as in the literature [66], indicating that the impact of pore water pressure on slope stability is almost negligible. Most other effect variables have important influences that cannot be ignored. Therefore, these factors are also an indispensable part of slope stability analysis. As different influence factor rankings may be obtained when different data sets and classification models are used [67], more representative results may be obtained with the emergence of more effective slope cases and algorithms in the future.

Conclusions
This research, based on the machine learning algorithm, employed the intellig gorithm optimization technique to construct an integrated classification machine, d the slope into stable and unstable, comprehensively considered the slope failure m nism and geological conditions, and collected 117 slope cases to establish a data s GA model with the biological evolution principle was used to optimize the hyperp eters of five machine learning methods: the RF, DT, SVM, KNN, and GBM models,

Conclusions
This research, based on the machine learning algorithm, employed the intelligent algorithm optimization technique to construct an integrated classification machine, divided the slope into stable and unstable, comprehensively considered the slope failure mechanism and geological conditions, and collected 117 slope cases to establish a data set. The GA model with the biological evolution principle was used to optimize the hyperparameters of five machine learning methods: the RF, DT, SVM, KNN, and GBM models, which were verified by the 10-fold CV method. The confusion matrix, ROC curve, and AUC value were used as evaluation indices to evaluate the accuracy of the model. The importance of the DT and GBM variables was analyzed, and the following conclusions were drawn.

1.
The adjustment result of hyperparameters by the genetic algorithm is relatively stable. The AUC value of the eight optimization results has little fluctuation, and the AUC value of the algorithm performs well. The average AUC value is between 0.858 and 0.956, which shows that the algorithm can effectively optimize the hyperparameters of the machine learning algorithm. In the test set results, the SVM model has the maximum AUC value (0.964), followed by RF (0.944) and DT (0.901) models, which shows that these three algorithms have good performance in slope stability prediction.

2.
The results of the 50% baseline cutoff and KS value cutoff of the K-S curve show that the slope stability prediction model with KS value cutoff generally has a better effect.
Considering the practicality of slope engineering problems, AUC value, ACC value, and TNR value, the RF model with a KS value cutoff point is recommended for slope stability prediction.

3.
The findings indicate that for the data set and algorithm in this study, the cohesion variable has the highest impact on slope stability prediction, with an influencing factor of 0.327, which accounts for about one-third of the total influence. Pore water pressure is the least obvious factor.