Risk Prediction of Coal and Gas Outburst in Deep Coal Mines Based on the SAPSO-ELM Algorithm

Effective risk prevention and management in deep coal mines can reduce the occurrences of outburst accidents and casualties. To address the low accuracy and inefficiency of coal–gas outburst prediction in deep coal mines, this study proposes a deep coal–gas outburst risk prediction method based on kernal principal component analysis (KPCA) and an improved extreme learning machine (SAPSO-ELM) algorithm. Firstly, high-dimensional nonlinear raw data were processed by KPCA. Secondly, the extracted sequence of outburst-causing indicator principal components were used as the input variables for the simulated annealing particle swarm algorithm (SAPSO), which was proposed to optimize the input layer weights and implied layer thresholds of the ELM. Finally, a coal and gas outburst risk prediction model for a deep coal mine based on the SAPSO-ELM algorithm was developed. The research results show that, compared with the ELM and PSO-ELM algorithms, the SAPSO-ELM optimization algorithm significantly improved the accuracy of risk prediction for coal–gas outbursts in deep coal mines, and the accuracy rate was as high as 100%. This study enriches the theory and methods of safety management in deep coal mines, and effectively helps coal mine enterprises in improving their ability to manage coal–gas outburst risks.


Introduction
Coal and gas outbursts are essentially caused by a process of stress concentration, strength destruction, sudden destabilization, and rapid ejection under the influence of gas coupled with a coal body and gas with adsorption properties [1,2]; it is a significantly complex natural disaster [3]. As the world's top coal-producing and consuming country, China was at 56% of its primary energy consumption in 2021. Its natural resources are "rich in coal, poor in oil, and low in gas", making it one of the most concerned countries in the world regarding prevention of coal-gas outburst [4][5][6]. The coal-mining depth is increasing at a rate of 10-25 m/a, owing to the gradual depletion of shallow coal resources. Based on statistics, the current deep (below 1000 m) coal resources constitute approximately 53% of the overall coal resources; there are 47 mines deeper than 1000 m that have a maximum mining depth of 1500 m, and more than 30 new 1000 m-deep mines will be built in the next 5-10 years [7]. Additionally, the geological conditions of deep coal seams in China are extremely complex, which threatens the safe coal mine production and safety of miners [8,9]. Therefore, the accurate prediction of coal-gas outbursts in deep coal mines is of great significance for ensuring the safety of deep-seated mine operations [10].
Many scholars globally have launched numerous studies on the issue of coal-gas outburst risk predictions. Acoustic emission and microseismic-monitoring methods are considered the two most common dynamic continuous methods that can effectively predict coal-gas outbursts [11][12][13]. Liu et al. [14] combined a gradient boosting decision tree with

Kernel Principal Component Analysis (KPCA)
KPCA [36,37] is mainly used to address nonlinear complex problems. It essentially employs a nonlinear function with extremely high nonlinear-processing capability, through which the introduced function maps a large amount of complex information in the original space to linearly divisible high-dimensional data. This function is widely used in fields such as dimensionality reduction in indicator features to extract the main information features, reduce information redundancy, and minimize the number of operations. In practical problem processing, Gaussian radial basis kernel functions are favored by scholars owing to their superior processing performance and outstanding anti-interference ability [38]. X = [x 1 , x 2 , · · · , x n ] T ∈ R n * m is the initial dataset, where n represents the total number of samples, and m represents the data dimensionality. The corresponding operation can be completed for the data points x 1 , x 2 , · · · , x n to transform them into a certain feature space F using the nonlinear function Φ. At this point, Φ(x k ) can be used for representation, and the covariance matrix formed within F is shown in Equation (1): If the eigenvalues and eigenvectors of C F are expressed in terms of λ and v, respectively, their relationship can be expressed by Equation (2): In Equation (2), λ = [λ 1 , λ 2 , . . . , λ N ] is an eigenvalue matrix obtained by solving for C F with each eigenvalue satisfying λ 1 ≥ λ 2 ≥, . . . , ≥ λ N , and α i (i = 1, 2, . . . M) represents the existence coefficient of a term, and this coefficient can provide some basis for the establishment of Equation (3): Substituting this into Equation (1) and multiplying Φ(x k ) on both sides of Equation (1) yields: A simplified operation of Equation (4) yields: In Equation (5) , and α = [α 1 , α 2 , . . . , α n ] T represents the associated eigenvectors in the kernel matrix K.
Before performing KPCA, matrix K is centralized by replacing K in Equation (5): In Equation (6), 1 N represents a type of N × N matrix in which each factor is 1/N, and N is the dimension of K.
Through Equations (5) and (6), it is possible to complete the calculation of the eigenvector v of matrix C F based on the eigenvector α of the K matrix. Subsequently, the cumulative variance contribution ratio generated by the relevant eigenvalues is calculated using Equation (7), and the top P eigenvalues are used to study based on the contribution ratio requirement. Simultaneously, the determination of the total number of principal components by data feature extraction can be realized based on these eigenvalues.

Extreme Learning Machine Prediction Model
The extreme learning machine (ELM) is a new single hidden layer feedforward neural network that can sufficiently handle multimodal classification problems such as deep coal mine gas outburst prediction. It maintains many advantages, such as high learning accuracy of general neural networks that improves the learning efficiency to a great extent, optimizes the gradient descent-based learning, straightforward computational process, strong generalization power, and it has been popularized in several industries and disciplines [39][40][41].
The forward single hidden layer architecture in ELM is chosen as the network training model. The number of nodes in the three parts of the network, such as the input layer, hidden layer, and output layer, are represented by m, M, and n, respectively; g(x) represents the activation function involved in the hidden layer neurons, and bi is the critical value. Assuming that the number of samples (x i ,t i ) with differentiation is N, 1 ≤ I ≤ N, the ELM structural model is as shown in Figure 1.
vector v of matrix C F based on the eigenvector α of the K matrix. Subsequently, the cumulative variance contribution ratio generated by the relevant eigenvalues is calculated using Equation (7), and the top P eigenvalues are used to study based on the contribution ratio requirement. Simultaneously, the determination of the total number of principal components by data feature extraction can be realized based on these eigenvalues.

Extreme Learning Machine Prediction Model
The extreme learning machine (ELM) is a new single hidden layer feedforward neural network that can sufficiently handle multimodal classification problems such as deep coal mine gas outburst prediction. It maintains many advantages, such as high learning accuracy of general neural networks that improves the learning efficiency to a great extent, optimizes the gradient descent-based learning, straightforward computational process, strong generalization power, and it has been popularized in several industries and disciplines [39][40][41].
The forward single hidden layer architecture in ELM is chosen as the network training model. The number of nodes in the three parts of the network, such as the input layer, hidden layer, and output layer, are represented by m, M, and n, respectively; g(x) represents the activation function involved in the hidden layer neurons, and bi is the critical value. Assuming that the number of samples (xi,ti) with differentiation is N, 1 ≤ I ≤ N, the ELM structural model is as shown in Figure 1.  The ELM structural model is mathematically expressed as follows: In Equation (8), , ω i2 , ω i3 , . . . , ω im ] represents the input weight vector that connects both the input layer node and i-th hidden layer node, β i = [β i1 , β i2 , β i3 , . . . , β in ] T represents the output weight vector that connects the output layer to the hidden layer nodes, and o i = [o i1 , o i2 , o i3 , . . . , o in ] T represents the network output value.
For the extreme learning machine, the cost function E can be expressed through the following equation: In Equation (9), S = (w i , b i , i = 1, 2, . . . , M) involves the input weights and hidden node critical values. The learning objective of the ELM computational model is to search for the ideal S and β, such that the error between the output and actual results is as small as possible, that is, min(E(S,β)).
min(E(S, β)) = min In Equation (10), H, β, and T represent the output matrix of the hidden layer, β output weight matrix, and target value matrix, respectively, and the formula defining H, β, and T can be expressed by the following equations: The ELM computational model essentially involves the process of solving the ideal solution of a nonlinearly varying target object. During this period, the process can be transformed into a least-squares solution-related computation for the minimum parametrization of the linear system Hβ = T:β = H + T In Equation (13), H + represents the MP generalized inverse of matrix H. This is the ELM training process.

Simulated Annealing Particle Swarm Algorithm (SAPSO)
Particle swarm optimization (PSO) is a swarm intelligence optimization algorithm proposed by American scholars James Kennedy and Russell Eberhart in 1995 [42,43], which focuses on the cooperation among particles by simply adjusting the behavior patterns of individual particles to achieve optimal solutions to complex problems. Particle swarm algorithms have the advantages of few adjustment variables, fast operation, and high accuracy, and are used in many fields, including optimization functions and machine learning [44]. However, it also has the disadvantages of not sufficiently converging quickly, and is prone to premature convergence; this study uses simulated annealing (SA) to improve it and obtain the simulated annealing particle swarm algorithm (SAPSO), as shown in Figure 2.
The SA algorithm [45,46], in seeking the optimal solution to the target problem, first determines the initial temperature and interferes with it, simulates the energy state transfer process of the particle at different temperatures, compares the new solution with the current key, and replaces it based on the Metropolis criterion. The SA algorithm accepts the optimal solution with probability 1 and agrees with the worst solution with some possibility as a way to escape the local optimal solution trap. The Metropolis criterion defines the probability of internal energy formation when an object is in a specific temperature T environment, and gradually transforms from states i to j, which is expressed by the following equation: In Equation (14), E(i) and E(j) represent the internal energy generated by the solid in two states, i and j. In turn, ∆E, K, and T represent the increment in internal energy, Boltzmann's constant, and temperature of the material itself, respectively. The SA algorithm [45,46], in seeking the optimal solution to the target problem, first determines the initial temperature and interferes with it, simulates the energy state transfer process of the particle at different temperatures, compares the new solution with the current key, and replaces it based on the Metropolis criterion. The SA algorithm accepts the optimal solution with probability 1 and agrees with the worst solution with some possibility as a way to escape the local optimal solution trap. The Metropolis criterion defines the probability of internal energy formation when an object is in a specific temperature T environment, and gradually transforms from states i to j, which is expressed by the following equation: In Equation (14), E(i) and E(j) represent the internal energy generated by the solid in two states, i and j. In turn, ΔE, K, and T represent the increment in internal energy, Boltzmann's constant, and temperature of the material itself, respectively.
By analyzing the advantages and disadvantages of the two algorithms from a comprehensive perspective, a PSO algorithm based on simulated annealing is created, which By analyzing the advantages and disadvantages of the two algorithms from a comprehensive perspective, a PSO algorithm based on simulated annealing is created, which enhances the particle swarm diversity, effectively suppresses the occurrence of premature convergence, and retains the advantages of the short local convergence time of the primary particle swarm algorithm, which can provide a sufficient guarantee for the accuracy of local search while improving the global search capability of the algorithm. The parameters of the SA algorithm were set as follows: the starting temperature was set to 1000 • C and the decay factor was 0.2. The parameters of the PSO algorithm were set as follows: the maximum number of runs was 100; the overall size of the species was 20; C1 and C2 were set to 2; and the inertia weight was set to 0.9. The SA algorithm can quickly determine the location of the solution in a short period and use it as its initial coordinates. It then uses the PSO algorithm for targeted screening to determine the optimal solution.

Construction of SAPSO-ELM Coal and Gas Outburst Prediction Model in Deep Coal Mines
When the ELM model is trained, its input layer weights and implied critical layer values are usually set unexpectedly and are not sufficiently stable. The built model is prone to several deficiencies, such as low prediction efficiency and weak generalization ability. Therefore, this study used the SAPSO algorithm to improve the ELM input layer weights and implied critical layer values to construct a SAPSO-ELM prediction model for coal-gas outburst prediction in deep coal mines. The steps of the optimization search were as follows.
Step 1: Randomly assign the training and test sets and normalize the data.
Step 2: Construct the network structure of the extreme learning machine, and then set the population size and number of particles.
Step 3: Set the parameters of the SAPSO algorithm.
Step 4: Record the optimal value of the current iteration.
Step 5: Determine when the error reaches the allowable range and when the number of runs reaches the idealized state. Stop the run when the ideal error disappears and start Step (7), and vice versa for Step (6).
Step 6: Update the particle velocity and position for iteration.
Step 7: Determine the optimal weights and thresholds.
Step 8: Train the whole test set and start the prediction.

Data Pre-Processing
In all the predictive indexes of coal-gas outbursts in deep coal mines, discretization between the magnitudes and orders of importance is produced. If the initial data are directly used, it will severely impact the stability and continuity of the model. Therefore, when studying the predictive indexes by KPCA, the normalized disposition operation of the initial data should be completed first, and the maximum-minimum method is selected for processing, with the specific formula as follows: In the case of positively correlated indicators, also called benefit-based indicators, the following treatment is performed: In the case of negatively correlated indicators, also referred to as cost-based indicators, the following treatment is performed: In the above equations, x max and x min represent, in turn, the two extreme values of a certain influence factor in the data information series embedded within the prominent predictor system, x i represents the predictor obtained by normalization, and x i represents the original indicator.

Feature Extraction
After normalizing the original data, we selected the Gaussian radial basis kernel function with good nonlinear exploration ability for correlation analysis, where σ represents the hyperparameter of the kernel function, and was set to 500 based on relevant experience [38], and the eigenvalues of all the predictors with their contribution rates can be obtained through certain processing; the specific data are shown in Table 2. It is shown that the cumulative contribution rate of the first five indicators is 91.17%, according to the principal component extraction criterion, when the cumulative contribution rate of the extracted main components reaches 90% or more, they can represent all the information contained in the principal components, which can effectively solve the degradation problem of prediction accuracy caused by high-dimensional data input, and significantly improve the computational speed and prediction accuracy of the prediction model. Therefore, the first five indicators were used in this study to replace the initial 12 predictors, that is, the input variables were set as Y 1 , Y 2 , Y 3 , Y 4 , and Y 5 in the model. The relevant data for some new variables after the feature extraction operation are listed in Table 3.

Training SAPSO-ELM Model Network Performance
The 178 initial datasets were randomly divided into 50 test sets and 128 training sets. ELM can achieve the learning purpose by adding only the number of hidden layers; the correspondence between the number of SAPSO-ELM hidden layer nodes and the test accuracy of the training set is shown in Table 4. For 8-35 hidden layer nodes, when the number of hidden layers reached 34, the test accuracy of the corresponding training set reached the highest value of 95.31%. Based on relevant experience, when the sigmoidal function was selected for the activation function, its generalization performance was the best, and the training error was the smallest [47]. Therefore, the author determined that the number of nodes in the implicit layer of the extreme learning machine was 34, and the network structure of the revolutionary learning machine was 5-34-4.

Comparison and Analysis of Prediction Results of Different Algorithms
After training the improved ELM model using the training samples, the degree of coal-gas outburst prediction of the test samples can be divided into four levels, based on the degree of damage produced by the coal-gas outburst phenomenon, which are indicated by 1, 2, 3, and 4. Thus, the overall distribution of the output image of the expected value is mainly point-like. The final output results may differ slightly if the system completes the prediction operation using multiple algorithm models, even for the same verification samples. Then, the comparison is performed between the difference and real data of the test samples to clearly evaluate the advantages and disadvantages of the algorithm. This test set was compared with the ELM and PSO-ELM algorithms to compare the prediction effect of each. The simulation results are shown in Figure 3; the accuracy of the SAPSO-ELM prediction model improved to 100%.

Comparison and Analysis of Prediction Results of Different Algorithms
After training the improved ELM model using the training samples, the degree of coal-gas outburst prediction of the test samples can be divided into four levels, based on the degree of damage produced by the coal-gas outburst phenomenon, which are indicated by 1, 2, 3, and 4. Thus, the overall distribution of the output image of the expected value is mainly point-like. The final output results may differ slightly if the system completes the prediction operation using multiple algorithm models, even for the same verification samples. Then, the comparison is performed between the difference and real data of the test samples to clearly evaluate the advantages and disadvantages of the algorithm. This test set was compared with the ELM and PSO-ELM algorithms to compare the prediction effect of each. The simulation results are shown in Figure 3; the accuracy of the SAPSO-ELM prediction model improved to 100%.

Comparison of SAPSO-ELM and ELM Prediction Results
To fully reflect the rationality of the ELM based on the improvement of the intelligent algorithm, this study was mainly simulated using MATLAB software, which is a commercial mathematics software produced by MathWorks of the United States. The ELM optimized without the intelligent algorithm was compared and analyzed with the ELM optimized using the SAPSO algorithm to verify its optimization effect. As shown in Figures 3 and 4, when tested on 50 random samples, the optimized extreme-learning machine model showed no misclassification, and the accuracy rate was 100%. The model without

Comparison of SAPSO-ELM and ELM Prediction Results
To fully reflect the rationality of the ELM based on the improvement of the intelligent algorithm, this study was mainly simulated using MATLAB software, which is a commercial mathematics software produced by MathWorks of the United States. The ELM optimized without the intelligent algorithm was compared and analyzed with the ELM optimized using the SAPSO algorithm to verify its optimization effect. As shown in Figures 3 and 4, when tested on 50 random samples, the optimized extreme-learning machine model showed no misclassification, and the accuracy rate was 100%. The model without optimization showed 10 misclassifications, with an accuracy rate of only 80%. Therefore, the prediction accuracy of the former was significantly higher than that of the other two. The biggest difference between these two algorithms is whether the SAPSO algorithm is used for the optimization. Compared with the unoptimized model, the method effectively overcomes the drawback of randomly selecting input weights and implicit threshold parameters in ELM. The accuracy of the model was significantly improved after the optimization operation, and it exhibited good generalization performance. Therefore, the improved ELM classification model was superior to the traditional ELM classification model.

Comparison of SAPSO-ELM and PSO-ELM Prediction Results
As shown in Figures 3 and 5, during the test, the SAPSO-ELM model did not produce misclassification and demonstrated an accuracy of 100%, whereas the PSO-ELM model produced four misclassifications with an accuracy of 92%. When predicting the 3rd, 32nd, 34th, and 35th datasets, the PSO-ELM predicts incorrect results and deviates from the actual results, while the SAPSO-ELM model can still predict the results correctly. This is because the particle swarm algorithm is more prone to the phenomenon of local optimum, which in turn generates premature convergence, and the simulated annealing particle swarm algorithm makes the particles deviate from the local optimum. Therefore, compared to the PSO-ELM model, the SAPSO-ELM model is more suitable for the required prediction.

Comparison of SAPSO-ELM and PSO-ELM Prediction Results
As shown in Figures 3 and 5, during the test, the SAPSO-ELM model did not produce misclassification and demonstrated an accuracy of 100%, whereas the PSO-ELM model produced four misclassifications with an accuracy of 92%. When predicting the 3rd, 32nd, 34th, and 35th datasets, the PSO-ELM predicts incorrect results and deviates from the actual results, while the SAPSO-ELM model can still predict the results correctly. This is because the particle swarm algorithm is more prone to the phenomenon of local optimum, which in turn generates premature convergence, and the simulated annealing particle swarm algorithm makes the particles deviate from the local optimum. Therefore, compared to the PSO-ELM model, the SAPSO-ELM model is more suitable for the required prediction.

Comparison of Comprehensive Forecast Results
First, in terms of the time cost for model fitting, the ELM network training model adopts a forward single hidden layer structure; therefore, the joint weights w for the interaction between the input layer and the hidden layer, and the adjustment bias b for the hidden layer are obtained by initialization. This process is stochastic and requires no adjustment during the training process, however, only the number of neurons in the hidden layer should be set, and the output of the learning network can be determined by a onestep calculation; hence, the standard model is trained in a randomized manner. In this study, the standard ELM takes the shortest time (0.2175 s) superimposed on the PSO algorithm rules in the model training; this is because the particle swarm algorithm requires fewer adjustment parameter selection with more mature theoretical support and it is easy to implement. The training time of PSO-ELM is 3.3555 s, which is higher than the standard ELM. In the embedded SA computation mode, it is necessary to continuously perform optimization, judgment, and replacement among the training data. Therefore, the SAPSO-ELM algorithm model takes the most time for this prediction, reaching 4.3872 s.
Second, in the SAPSO-ELM algorithm, the ELM weights and thresholds are randomly generated, which may affect the accuracy of coal-gas outburst evaluation in deep coal mines. The PSO algorithm easily falls into local optimal solutions and produces premature convergence, and the most suitable ELM weights and thresholds are found by the SAPSO algorithm, such that the model generalization ability is enhanced to some extent.
As seen in Table 5, the SAPSO-ELM algorithm has a stronger generalization ability in both training and prediction accuracies than the unoptimized ELM and PSO-ELM prediction models. In terms of prediction time, the prediction speed of all three models was improved to different degrees after KPCA was adopted to extract the feature metrics. Although the SAPSO-ELM model requires a slightly longer time in the prediction phase than the other models, the accuracy and stability are more important in highlighting the prediction effect if the prediction time is within the specified interval. Overall, the advantages of the prediction model based on KPCA and SAPSO-ELM proposed in this study are more

Comparison of Comprehensive Forecast Results
First, in terms of the time cost for model fitting, the ELM network training model adopts a forward single hidden layer structure; therefore, the joint weights w for the interaction between the input layer and the hidden layer, and the adjustment bias b for the hidden layer are obtained by initialization. This process is stochastic and requires no adjustment during the training process, however, only the number of neurons in the hidden layer should be set, and the output of the learning network can be determined by a one-step calculation; hence, the standard model is trained in a randomized manner. In this study, the standard ELM takes the shortest time (0.2175 s) superimposed on the PSO algorithm rules in the model training; this is because the particle swarm algorithm requires fewer adjustment parameter selection with more mature theoretical support and it is easy to implement. The training time of PSO-ELM is 3.3555 s, which is higher than the standard ELM. In the embedded SA computation mode, it is necessary to continuously perform optimization, judgment, and replacement among the training data. Therefore, the SAPSO-ELM algorithm model takes the most time for this prediction, reaching 4.3872 s.
Second, in the SAPSO-ELM algorithm, the ELM weights and thresholds are randomly generated, which may affect the accuracy of coal-gas outburst evaluation in deep coal mines. The PSO algorithm easily falls into local optimal solutions and produces premature convergence, and the most suitable ELM weights and thresholds are found by the SAPSO algorithm, such that the model generalization ability is enhanced to some extent.
As seen in Table 5, the SAPSO-ELM algorithm has a stronger generalization ability in both training and prediction accuracies than the unoptimized ELM and PSO-ELM prediction models. In terms of prediction time, the prediction speed of all three models was improved to different degrees after KPCA was adopted to extract the feature metrics. Although the SAPSO-ELM model requires a slightly longer time in the prediction phase than the other models, the accuracy and stability are more important in highlighting the prediction effect if the prediction time is within the specified interval. Overall, the advantages of the prediction model based on KPCA and SAPSO-ELM proposed in this study are more evident, mainly in the small training error, strong generalization ability, and good prediction accuracy, which can play a vital role in preventing and controlling coal-gas outburst disasters in deep coal mines when applied.

Testing the Effect of SAPSO-ELM Prediction Model
To verify the generalization performance of the SAPSO-ELM prediction model, the ROC curve, also known as the receiver operating characteristic curve [48], was used to evaluate the classifier validity, with values closer to the upper left corner of the curve indicating a better classification effect. When it is difficult to judge, it can be measured by the quantitative area under the curve (AUC) metric, with larger values indicating better classification effectiveness, in the range of 0.5 and 1 [49]. Gas outbursts are classified into four categories based on their outburst size, which is essentially a data classification problem, and a confusion matrix is introduced to evaluate the dichotomous classification problem in Table 6. True/False represents whether the classification is correct, and Positive/Negative represents the category label of the predicted sample. Here, five indicators, SEN, SPE, G-mean, F-measure, and AUC, are selected to test the validity of the SAPSO-ELM prediction model, and the solution formula is as follows: As shown in Figure 6, the classification curve of ELM is closest to the diagonal, its corresponding AUC is 0.87, and the corresponding AUC of PSO-ELM is 0.95. The SAPSO-ELM model proposed in this study corresponds to the largest AUC of up to one, which is in the ideal classification state. As presented in Table 7, the five metrics obtained by the calculation show that SAPSO-ELM is generally higher than the other two prediction models. Therefore, the improved ELM with the SAPSO is effective in improving the classifier performance.

Discussions
With the depletion of shallow coal resources, the mining intensity and depth of coal mines are increasing gradually. Moreover, the geological conditions of deep coal seams are becoming incredibly complex; the coal seam inclination changes significantly with many folds and faults. Under the combined effect of soft coal, low coal permeability, high stress, high gas content, and high gas pressure, coal seams show more complicated characteristics that lead to an increase in frequency and intensity of coal mine disasters. This also induces accidents, such as rockbursts, and coal and gas outbursts, posing a severe threat to mine safety. The aforementioned factors have led to new characteristics of gas outbursts in deep coal mines. In addition, the coupling between multiple hazards has led to a more complex mechanism of gas outbursts and requires an in-depth analysis of the conditions. The mechanism of gas outburst is an important prerequisite for effective prevention of this hazard risk. To further improve the accuracy of coal-gas outburst prediction in deep coal mines, this study improved the original gas outburst prediction index system based on an in-depth analysis of coal-gas outburst mechanisms and influencing factors, combined with examples of gas outburst accidents in deep coal mines and the existing literature on outburst prediction indices. A coal-gas outburst prediction model

Discussions
With the depletion of shallow coal resources, the mining intensity and depth of coal mines are increasing gradually. Moreover, the geological conditions of deep coal seams are becoming incredibly complex; the coal seam inclination changes significantly with many folds and faults. Under the combined effect of soft coal, low coal permeability, high stress, high gas content, and high gas pressure, coal seams show more complicated characteristics that lead to an increase in frequency and intensity of coal mine disasters. This also induces accidents, such as rockbursts, and coal and gas outbursts, posing a severe threat to mine safety. The aforementioned factors have led to new characteristics of gas outbursts in deep coal mines. In addition, the coupling between multiple hazards has led to a more complex mechanism of gas outbursts and requires an in-depth analysis of the conditions. The mechanism of gas outburst is an important prerequisite for effective prevention of this hazard risk. To further improve the accuracy of coal-gas outburst prediction in deep coal mines, this study improved the original gas outburst prediction index system based on an in-depth analysis of coal-gas outburst mechanisms and influencing factors, combined with examples of gas outburst accidents in deep coal mines and the existing literature on outburst prediction indices. A coal-gas outburst prediction model for deep coal mines was constructed based on KPCA and SAPSO-ELM algorithms. The prediction model was also applied to predict coal-gas outbursts from mine excavation studies, which provided a significant reference value for improving the safety of deep mine production.
Most of the existing literature contains information on shallow coal mining. Studies have been conducted on some of the influencing factors, however, these studies mostly ignore the complicated nonlinearity and correlation between these factors, which significantly affect the reliability of predicting gas outbursts in deep coal mines. With the continuous extension of coal mining depth, it is necessary to construct green, efficient, safe, and intelligent mines, and promote the process of neutralizing the carbon peaks; thus, it is of practical and theoretical significance to predict coal-gas outbursts in deep coal mines. In this study, we analyzed the coal-gas outburst hypothesis, considered the data accessibility to systematically understand the outburst-causing factors, and used the improved intelligent algorithm to make a scientific prediction. The example proved that the proposed model exhibits a good generalization performance, and can effectively prevent and address gas outburst disasters, which can guide coal mining enterprises to perform safe and effective deep coal mining.
Furthermore, with the increasing depth and intensity of coal seam mining, the geological conditions become more complex, coupling complexity between multiple disasters becomes higher, and corresponding influencing factors increase significantly. Therefore, in future research, it is necessary to widen the research on the process of coal-gas outbursts and related mechanisms in deep coal mining. Moreover, determining the coupling relationship between multiple coal-mining disasters and increasing the application of intelligent equipment in monitoring deep coal mine power disasters through numerous channels will further aid in continuous improvement in the management of coal mine enterprises, effective prevention and control of coal-gas outburst disasters, and realize safe and efficient mining in deep coal mines.

Conclusions
(1) In this study, we improved the original gas outburst prediction system by combining the examples of gas-outburst accidents in deep coal mines and the existing literature on gas outburst prediction indices, and constructed 12 deep mine gas outburst prediction indices, including coal seam mining depth, coal thickness variation coefficient, coal body damage type, coal seam lithology, mine temperature, soft partition thickness, gas pressure, coal seam gas content, coal seam resolvable gas content, coal seam permeability coefficient, coal body solidity coefficient, and gas discharge initial velocity. The 12 deep mine gas outburst prediction indices were analyzed by KPCA, and then reduced in dimensionality to solve the problem of input variables of the complex machine learning model and significantly improve the computational rate of the prediction model.
(2) We used the PSO algorithm to improve the ELM. Simultaneously, given the shortcomings of the PSO algorithm in terms of slow convergence speed and premature convergence, a prediction model of deep coal mine gas outburst based on SAPSO of the ELM was created. It fundamentally realizes the optimization of the prediction model of the ELM, enhances the generalization performance and stability of the model, and significantly improves the accuracy of gas outburst prediction in deep mines.
(3) The prediction accuracies of the SAPSO-ELM, PSO-ELM, and ELM models were 100%, 92%, and 80%, respectively, with training times of 4.3872 s, 3.3555 s, and 0.2175 s, respectively. We used MATLAB software to conduct relevant simulation tests and comparative studies on actual data of gas outbursts in deep coal mines. It was shown that the SAPSO-ELM model has the highest validity in determining the optimal global solution and can predict the gas outburst intensity of deep mines more accurately. In addition, although the SAPSO-ELM model takes slightly more time than the other models in the prediction process, its accuracy and stability are more critical in the outburst prediction effect if the prediction time is within a specified interval. Therefore, the gas outburst prediction model for deep coal mines proposed in this study has better generalization performance, and it analyzes and predicts deep coal mine gas outbursts from a new perspective, which is beneficial for ensuring safety in deep mines.
(4) The ROC curve was used to further test the proposed coal-gas outburst prediction model for deep coal mines. Its evaluation indices, such as SEN, SPE, G-mean, AUC, and F-measure, were significantly higher than those of other prediction models, which proved that the proposed prediction model has a good generalization ability. Its advantages are mainly reflected in its higher prediction accuracy and faster operation speed, which can be used for disaster prevention and control in deep coal mines.