1. Introduction
Coal continues to play a pivotal role in China’s energy consumption, acting as a primary energy source in critical sectors such as electricity generation, industry, and transportation, owing to its abundant reserves and relatively low cost [
1,
2,
3]. However, as mining depth increases, the complexity of gas emissions also rises, posing significant safety risks in underground operations. Gas emissions not only deteriorate the mine environment and disrupt equipment function, but they also elevate underground gas concentrations, substantially increasing the risk of gas explosions and presenting a direct threat to the continuity of mining activities and miner safety. The variability and unpredictability of gas emissions, influenced by factors such as geological structures, coal seam characteristics, and mining methods, render traditional prediction approaches insufficiently responsive. Therefore, accurately forecasting gas emissions is crucial for ensuring mine safety and mitigating the occurrence of hazardous incidents [
4,
5,
6,
7,
8].
Extensive research has been conducted both domestically and internationally on methods for predicting gas emissions in mining. Guo et al. [
9] designed a gas emission prediction system using wavelet denoising combined with BP neural networks and validated its accuracy experimentally. Zhang et al. [
10] applied the Bootstrap sampling method, using out-of-bag (OOB) data scores (oob_score) to tune the model and assess feature importance, ultimately determining the optimal parameters and feature variable importance proportions for constructing their predictive model. Yan et al. [
11] utilized factor analysis to extract prediction indicators and developed a gas emission prediction model using a particle swarm-optimized radial basis function (RBF) neural network. However, BP neural networks are prone to noise and outliers, with training processes that can become trapped in local optima; random forest (RF) models lack interpretability, making it challenging to analyze the specific contribution of each feature to the prediction results; and RBF models face issues like the “curse of dimensionality” when handling high-dimensional data, potentially leading to decreased model performance. Furthermore, these models tend to have limited generalization abilities, low robustness to small sample sizes, and difficulties in capturing complex nonlinear relationships, thereby constraining their precision in gas emission prediction [
12,
13,
14].
In this study, we address the complexity and multifactorial nature of gas emission prediction. Initially, data were preprocessed through cleaning and normalization, followed by the establishment of a comprehensive index system for gas emission prediction. To enhance predictive accuracy, Kernel Principal Component Analysis (KPCA) was introduced to perform feature extraction and dimensionality reduction, preserving critical information while reducing dimensional complexity, which improves model learning efficiency. Furthermore, an adaptive neighborhood search combined with reinforcement learning optimizes the Crow Search Algorithm (CSA) to fine-tune the SVR model’s hyperparameters, enhancing convergence speed and global search capability. The resulting KPCA-ICSA-SVR model effectively captures the nonlinear characteristics of gas emissions, yielding more accurate and reliable predictions [
15].
2. Formulation of Foundational Principles for Processes and Algorithms
2.1. Process for Establishing Predictive Models
The specific process for establishing a coal and gas outburst prediction model is as follows:
(1) Data Collection and Preprocessing: Gas emission data are collected and undergo comprehensive preprocessing. Outliers within the dataset are identified and removed using Boxplot analysis, thereby minimizing the impact of extreme values on model accuracy. Missing values are addressed through multiple imputation (MI), improving the completeness and quality of the data. Normalization is applied to remove scaling effects between features. Subsequently, Kernel Principal Component Analysis (KPCA) is employed for feature extraction and dimensionality reduction on the preprocessed data, retaining essential variable information while reducing dimensionality to optimize computational efficiency.
(2) Algorithm Optimization with Improved Crow Search (ICSA): An Improved Crow Search Algorithm (ICSA) is developed by integrating adaptive neighborhood search and reinforcement learning strategies. The adaptive neighborhood search adjusts search scopes dynamically, enhancing the algorithm’s flexibility, while the reinforcement learning strategy accumulates experience to improve optimization efficiency and accuracy throughout the process. The ICSA’s robust global and local search capabilities enable efficient exploration of the parameter space, fine-tuning the SVR model’s hyperparameters, and yielding an optimal parameter set for gas emission prediction.
(3) Model Construction and Comparative Evaluation: After data preprocessing and hyperparameter optimization, comparative models are established to evaluate prediction effectiveness. Benchmark models such as BP, KPCA-ICSA-BP, and SVR are constructed. Evaluation metrics, including RMSE and MPE, are used to assess the performance of each model in gas emission prediction. Results indicate that the KPCA-ICSA-SVR model demonstrates a substantial advantage in prediction accuracy and generalization capability, capturing the nonlinear characteristics of gas emissions more effectively.
2.2. Basic Principles of the Algorithm
2.2.1. Kernel Principal Component Analysis, KPCA
Principal Component Analysis (PCA) is a classical dimensionality reduction technique that projects data into a lower-dimensional space via linear transformations, aiming to preserve as much of the original data’s variance as possible. PCA is widely used in data preprocessing, feature extraction, and pattern recognition. However, the linear assumptions of PCA limit its effectiveness in handling complex, high-dimensional nonlinear data. In many real-world applications, data often exhibit significant nonlinearity, which constrains PCA’s capacity to fully capture the inherent patterns and relationships within the data [
16,
17].
To address this limitation, KPCA was developed [
18,
19]. By incorporating kernel techniques, KPCA maps data into a high-dimensional feature space where nonlinear features can be more effectively captured using kernel functions such as the Gaussian or polynomial kernel. This approach enables principal component analysis in this new feature space, retaining the dimensionality reduction advantages of PCA while significantly enhancing its ability to model nonlinear structures. KPCA has demonstrated superior performance in fields like pattern recognition and data mining, effectively improving both the accuracy and robustness of data analysis (
Figure 1).
(1) Data Standardization: The initial step in KPCA involves standardizing the original dataset X = {x
1,x
2,…,x
n} to ensure that all features are on a similar scale. For each sample, x
i∈R
m. The standardization formula is as follows (1):
where
µ is the mean and
σ is the standard deviation.
(2) Select Kernel Function: Choose an appropriate kernel function
K. In this study, the Gaussian kernel is selected, with the calculation formula given by the following (2):
(3) Construct Kernel Matrix: Calculate the kernel matrix K, where the elements are given by Kij = K(xi,xj). The dimension of the kernel matrix is n × n, where n is the number of samples.
(4) Center the Kernel Matrix: Perform centering on the kernel matrix to eliminate the influence of the mean. The centering formula is given by the following (3):
where 1
n is an
n ×
n matrix of ones.
(5) Compute Eigenvalues and Eigenvectors: Perform eigenvalue decomposition on the centered kernel matrix K
centered to obtain the eigenvalues
λi and the corresponding eigenvectors
vi (4):
(6) Select Principal Components: Choose the top d eigenvectors in descending order of eigenvalues to form the feature matrix Vd = [v1,v2,…,vd].
(7) Dimensionality Reduction Mapping: Map the original data into a lower-dimensional feature space to obtain the reduced data Y (5):
Typically, the reduced representation of each sample is expressed as follows (6):
where
αj is the coefficient of the corresponding eigenvector.
(8) Result Output: Output the reduced data Y, where each column represents a principal component.
2.2.2. Improved Crow Search Algorithm, ICSA
The Crow Search Algorithm (CSA) is a population-based intelligent optimization algorithm inspired by the foraging behavior of crows [
20,
21,
22]. By simulating the memory and defense strategies of crows, the CSA searches for optimal solutions within the solution space. It is characterized by its simplicity and wide applicability, having demonstrated effective results across various optimization problems. However, the CSA tends to get trapped in local optima during the search process, and its balance between exploration and exploitation is limited, leading to slower convergence speeds and insufficient precision in complex, multi-modal, high-dimensional problems [
23].
To address these limitations, this study proposes an Improved Crow Search Algorithm (ICSA), which integrates adaptive neighborhood search and reinforcement learning strategies. First, adaptive neighborhood search dynamically adjusts the search radius of crows, ensuring flexibility in exploration during different search phases; this allows for extensive searching in the early stages while gradually focusing on fine-tuning in later stages. Furthermore, the incorporation of reinforcement learning allows the crows to adaptively select optimal search behaviors based on feedback from the environment, thereby enhancing the algorithm’s decision-making capabilities. This improvement not only strengthens the CSA’s global exploration capabilities and local search accuracy but also significantly increases convergence speed and solution precision, resulting in enhanced robustness and stability in tackling complex optimization problems. A schematic representation of the ICSA is shown in
Figure 2.
The detailed steps and key formulas for the Improved Crow Search Algorithm (ICSA) are as follows:
(1) Initialize Population: Initialize the position xi and memory position mi of each crow in the population, along with their respective fitness values. Set the adaptive neighborhood radius δi and learning rate αi, as well as the Q-table for reinforcement learning, which will be used to learn the optimal behaviors for different states.
(2) Adaptive Neighborhood Search: Establish the neighborhood radius
δi and dynamically adjust the search neighborhood. Define the initial radius
δinit and a scaling factor
γ (where
γ ∈ (0, 1)). Adjust the neighborhood radius based on the current population density and fitness, allowing the search range to adaptively change during the optimization process (7):
where
t is the current iteration, and
γ controls the contraction rate.
(3) Reinforcement Learning Strategy: Apply Q-learning in reinforcement learning to establish a behavior selection strategy. Define the state
S (including information like current position and fitness) and the actions
A (choosing neighborhood search or extended search). Update each crow’s Q-value with the learning rate
η and discount factor
λ. The Q-value update formula is as follows (8):
where
r is the reward obtained from the current action, and
S’ is the new state resulting from action
A.
(4) Position Update: Based on neighborhood search and reinforcement learning decisions, choose whether to utilize the memory position (for local search) or perform global exploration (9):
where
ri is a random number in the range (0,1).
(5) Memory Position Update: If the new position Xit+1 is better than the memory position mi, update mi with Xit+1.
(6) Adaptive Adjustment of Neighborhood Radius and Learning Rate: Based on the crow’s fitness and population diversity, adaptively adjust the neighborhood radius
δi and learning rate
αi to ensure a larger exploration range in the early stages and more focused local searches later on (10):
where
β and
γ control the decay rates of the radius and learning rate, respectively, and
T is the maximum number of iterations.
(7) Stopping Criteria: The iteration process halts if a predefined maximum number of iterations is reached or if convergence criteria are met (e.g., the fitness value shows no significant improvement), and the optimal solution is outputted.
2.2.3. Support Vector Regression, SVR
Support Vector Regression (SVR) is a regression model well suited for scenarios with small sample sizes and high-dimensional data [
24,
25,
26]. In gas emission prediction, where data collection is challenging and sample sizes are typically limited, traditional methods may struggle to achieve satisfactory accuracy under small sample conditions or in high-dimensional spaces. SVR constructs an optimal regression hyperplane in the feature space, minimizing prediction error while controlling it within an
ϵ-insensitive margin, providing strong generalization ability and robustness to small sample sizes. This approach effectively handles the nonlinearity and high-risk factors inherent in gas emission prediction, making it ideal for small-sample data analysis and significantly enhancing prediction accuracy and reliability. An SVR illustration is shown in
Figure 3.
The detailed steps and key formulas for the SVR are as follows:
(1) Data Preparation: A training dataset is given by {(xi,yi)}ni=1, where xi∈Rm is the feature vector of the i-th sample, and yi∈R is the corresponding continuous target value.
(2) Kernel Trick: Since the data may not be linearly separable, a kernel function is used to map the data into a higher-dimensional space, making them closer to linearly separable in the new space. In this study, the Gaussian kernel is chosen.
(3) ϵ-Insensitive Loss and Soft Margin SVR: The
ϵ-insensitive loss function is introduced, allowing the model to ignore errors within the ±
ϵ range. This enables the model to tolerate some error within the
ϵ margin without penalizing the optimization objective. Slack variables
ξi ≥ 0 and
ξi* ≥ 0 are introduced to represent errors that exceed the
ϵ margin on the upper and lower sides, respectively. The objective is to minimize the complexity of the regression model while controlling the prediction error (11):
where
ξi and
ξi* represent the errors above the upper bound and below the lower bound, respectively.
C is a hyperparameter that balances the model complexity and the penalty for errors.
(4) Dual Problem: To incorporate the kernel function, the original optimization problem is transformed into its dual form. Lagrange multipliers
αi and
αi* are introduced to represent the weights for each data point. The dual problem can be expressed as follows (12):
subject to 0 ≤
αi,
αi* ≤
C and
.
(5) Support Vector Identification: After solving the dual problem, the Lagrange multipliers αi and αi* are obtained. For data points xi where αi > 0 or αi* > 0, these points are considered support vectors, which play a decisive role in determining the position of the regression hyperplane.
(6) Decision Function: After identifying the support vectors, the final decision function can be written as follows (13):
where
b is the bias term, which can be calculated using the support vectors.
(7) Regression Process: Use the obtained decision function f(x) to predict the continuous target values for new input data, evaluating the regression results based on these predicted values.
3. Data Preprocessing
3.1. Data Acquisition
This study uses data from a coal mine’s comprehensive mining face, where gas over-limit situations frequently occur. To prevent major disasters such as gas explosions, this research focuses on predicting gas outburst quantities in the mining face. A total of 30 representative data samples containing 12 influencing factors were obtained from the monitoring data of the mine to serve as the original dataset, as shown in
Table 1.
In
Table 1, the 12 influencing factors are as follows: coal seam burial depth
x1 (m), coal seam gas content
x2 (m
3/t), coal seam thickness
x3 (m), mining intensity
x4 (t/d), working face recovery rate
x5 (%), working face length
x6 (m), coal seam dip angle
x7 (°), adjacent layer gas content
x8 (m
3/t), adjacent layer thickness
x9 (m), interlayer distance
x10 (m), mining height
x11 (m), and advance speed
x12 (m/d); the gas outburst quantity is represented as
y (m
3/t). A total of 24 samples were randomly selected for the training set, while 6 samples were reserved for the testing set.
3.2. Data Cleaning Based on Boxplot–MI
Table 1 presents the uncleaned data. MATLAB was used to clean the original dataset, beginning with the identification and handling of outliers using the Boxplot method. This involved calculating the lower and upper quartiles (Q1 and Q3) and the interquartile range (IQR) for each column of data, establishing upper and lower limits to identify which data points were outliers. These outliers were replaced with NaN to prevent them from affecting data analysis in subsequent processing. Next, multiple imputation (MI) was applied to fill in these missing values. In this study, six estimators were employed for imputing the missing values, including linear regression, logistic regression, decision trees (DTs), random forests (RFs), a support vector machine (SVM), and multilayer perceptrons (MLPs). The root mean square error (RMSE) was used to evaluate the imputation results and determine the best estimator.
Figure 4 illustrates the RMSE for each estimator. Observations from
Figure 4 indicate that the MLP imputation model, which achieved the lowest RMSE, demonstrates the highest accuracy. MLP neural networks excel in handling classification and regression problems, showing strong predictive performance and robustness. They are particularly well suited for complex datasets and high-dimensional data. Therefore, this study selected an MLP as the estimator for MI data imputation. The complete data are presented in
Table 2 [
27,
28].
3.3. Data Dimensionality Reduction Based on KPCA
Kernel Principal Component Analysis (KPCA) utilizes kernel functions to map data into a high-dimensional feature space, effectively extracting underlying nonlinear structures and complex features. This characteristic significantly enhances the accuracy and efficiency of data analysis when dealing with high-dimensional and complex datasets. Furthermore, KPCA not only reduces dimensionality to improve the computational performance of models but also mitigates the impact of noise, thereby enhancing the expressive power of features.
Figure 5 shows the kernel matrix heatmap, which illustrates the similarity between data samples and reflects the structure of the kernel matrix calculated using a Gaussian kernel function. This heatmap provides a clearer analysis and identification of clustering trends and potential nonlinear relationships among the data samples. Each cell in the figure represents the similarity between two samples, with the color intensity indicating the level of similarity. Darker areas indicate high similarity, while lighter areas indicate low similarity. By observing
Figure 5, clustering trends and potential nonlinear relationships between the samples can be visually identified. For instance, the similarity between the data in Group 1 and the data in Groups 2 and 3 is relatively high, as shown by the darker areas in the heatmap, indicating that these three groups exhibit similar patterns or behaviors in the feature space and are likely to belong to the same or nearby categories. In contrast, the similarity between the data in Group 1 and the data in Groups 10, 14, and 15 is low, as indicated by the lighter areas, suggesting significant differences between these samples and implying that they likely belong to different categories.
The variance explanation graph is shown in
Figure 6, illustrating the proportion of total variance explained by each principal component. This graph serves as a tool for evaluating the importance and effectiveness of these components. The horizontal axis represents the component number, while the vertical axis shows the proportion of variance explained by each principal component. Typically, the first few principal components account for the majority of the variance, guiding the selection of an appropriate number of components to retain as much of the original data information as possible while reducing dimensionality and lowering computational complexity. Generally, a cumulative variance explanation ratio of 85% is considered a good balance between information retention and dimensionality reduction. In
Figure 6, the cumulative variance explained by the first four principal components is calculated as 0.629978 + 0.116409 + 0.080876 + 0.062901 = 0.890164, which is greater than 0.85. Therefore, the first four principal components are selected for the reduced-dimensional data, with the reduced dataset presented in
Table 3.
4. Establishment of Prediction Model
4.1. Optimization of Hyperparameters Based on the ICSA
Figure 7 illustrates the optimization capabilities of the improved and original algorithms on unimodal and multimodal test functions. As shown, the fitness values of the ICSA convergence curves on all test functions improved after incorporating adaptive neighborhood search and reinforcement learning strategies. This indicates that these strategies effectively enhance the global optimization capability of the CSA.
In this study, three hyperparameters are optimized: Penalty Parameter C: This parameter controls the model’s tolerance for error and determines the penalty for training errors. A higher C value makes the model focus more on reducing training errors, potentially leading to overfitting. Conversely, a lower C value yields a smoother model but may result in underfitting. Gaussian Kernel Bandwidth γ: When using the RBF (Radial Basis Function) kernel, γ controls the similarity measurement between samples and influences the model’s nonlinear mapping capability. A larger γ value allows the model to capture finer details, increasing flexibility but risking overfitting. A smaller γ value produces a smoother model with a higher risk of underfitting. Loss Function Parameter ϵ: This parameter in the ϵ-insensitive loss function defines the error tolerance range. A larger ϵ value allows the model to ignore minor errors, reducing sensitivity to noise. A smaller ϵ value improves model precision but may make it overly sensitive to noise.
After training, the optimal values for each parameter were obtained. The initialization range and optimal values of the ICSA parameter optimization are shown in
Table 4.
4.2. Establishment of the KPCA-ICSA-SVR Prediction Model
Substituting the optimal hyperparameter values obtained from the ICSA optimization into the SVR model, a gas emission prediction model based on KPCA-ICSA-SVR is established. The training set prediction results are shown in
Figure 8. As observed from the figure, the RMSE of the training set prediction reaches 0.17898, with an R
2 of 0.9836, MAE of 0.1651, MSE of 0.032, and MAPE of 1.67%. All performance metrics are within reasonable ranges, indicating that the KPCA-ICSA-SVR model developed in this study for gas emission prediction is both reasonable and feasible.
The test set prediction results are shown in
Figure 9. From the figure, it can be observed that the RMSE of the test set prediction reaches 0.3132, indicating that the model’s accuracy is within a reasonable range. This suggests that the KPCA-ICSA-SVR model does not exhibit overfitting or underfitting, demonstrating good generalization ability and enabling accurate predictions on new data.
5. Evaluation and Comparison of Models
5.1. Test Set Performance Evaluation
To validate the superiority of the KPCA-ICSA-SVR model established in this study, three comparative models were constructed: BP neural network [
29,
30,
31], KPCA-ICSA-BP, and SVR. Each model was applied to the test set for prediction, with the results presented in
Table 5. Due to the small dataset size, the computation time for all models is within 15 s. As shown in the table, the KPCA-ICSA-SVR model outperforms others across all evaluation metrics. Notably, the addition of KPCA for dimensionality reduction and the ICSA to the BP model significantly enhanced its predictive accuracy, indicating that the KPCA-ICSA method can effectively optimize prediction models.
(1) BP Model
The BP model performed moderately in predicting gas emissions, with an R2 value indicating that approximately 85% of the data variance could be explained, suggesting a degree of underfitting. Both MAE and RMSE were relatively high, reflecting the model’s limitations in minimizing error, particularly in responding to larger errors (as RMSE slightly exceeded MAE). The MAPE value showed considerable variation across different data magnitudes, with a positive bias (MPE = 4.5%), indicating that BP predictions tended to overestimate actual values. This highlights the model’s lower sensitivity to small samples and extreme values. An MBE value of 0.4 confirmed a notable positive bias, indicating that the BP model tends to overestimate gas emissions.
Upon incorporating KPCA for feature extraction and the ICSA to optimize the BP model’s hyperparameters, the R2 value improved significantly to 0.92, demonstrating that the model could explain 92% of data variance, with a considerable boost in fitting ability. Compared to the BP model, the KPCA-ICSA-BP model’s errors were notably reduced, reflecting that feature extraction and hyperparameter optimization successfully minimized predictive bias and improved stability. The model’s MAPE decreased by approximately 2.5%, enhancing prediction accuracy across various samples. A reduction in MPE indicated a better error distribution, and a lower MBE demonstrated decreased positive bias, confirming that KPCA and the ICSA effectively reduce systematic error.
(2) SVR Model
The SVR model outperformed the BP series in terms of fitting capability, achieving an R2 value of 0.94, indicating SVR’s advantage in capturing the nonlinear characteristics of gas emission data. Both MAE and RMSE were better than those of BP and KPCA-ICSA-BP, showing that SVR effectively manages larger errors. Lower overall error highlighted SVR’s accuracy in predicting gas emissions, with a MAPE of 3.5%, reflecting strong stability in predictions. The negative MPE value suggested a slight tendency to underestimate gas emissions, though the error magnitude was small. The negative MBE value also confirmed a mild underestimation tendency, with minimal error.
The KPCA-ICSA-optimized SVR model achieved the highest R2 value of 0.9752, demonstrating exceptional ability to explain the variance in gas emissions and superior fitting capability. Among the models, KPCA-ICSA-SVR had the lowest error rates, indicating robust control over both small and large errors. The MPE value was closer to zero, further affirming the model’s stability and precision in gas emission prediction. The near-zero MBE value suggested a very slight underestimation, but with minimal bias, making the model highly effective for prediction.
5.2. Model Safety Adjustment
Based on the model’s prediction results, a safety margin is added to the predicted gas emission values (14):
where Δ
safety is a safety adjustment value set based on historical data and expert experience to ensure early warnings under high-risk conditions. The calculation formula is as follows (15):
where
α is a compensation coefficient determined by the historical extreme values of gas emissions and the risk tolerance level. To ensure sufficient conservativeness, this study sets
α = 0.05, resulting in a Δ
safety value of 0.165.
After incorporating the safety margin, the Mean Bias Error (MBE) in the model evaluation metrics reaches a positive value of 0.05, effectively improving the model’s safety.
5.3. Model Generalization Verification
To evaluate the applicability of the proposed model, 30 sets of field data from the reference [
11] were used for validation. In [
11], the test mine, located at the 12,322 main working face, has an average coal seam thickness of 6.25 m, including one to two layers of carbonaceous mudstone and mudstone interlayers. In areas affected by geological structures, the coal seam thickness varies significantly, with faults extending into the working face. These factors—variations in coal seam thickness, faulting, and interlayer disruptions—can cause periodic fluctuations in gas emission. Moreover, fault damage and stress concentration zones may lead to sudden large gas emissions, increasing the risk of unexpected outbursts. Therefore, accurate gas emission prediction is essential to ensure safe and orderly mining operations.
The sample data include 11 indicators, such as daily production of the working face, interlayer lithology, and coal seam thickness. The data samples were split into two groups; 80% were used for training and 20% for testing. The performance prediction results of different models are summarized in
Table 6.
As shown in the table, the KPCA-ICSA-SVR gas emission prediction model established in this study outperforms the models referenced in the literature across all evaluation metrics. This demonstrates that the proposed model has good applicability to other coal seams.
6. Discussion
In summary, the gas emission prediction model based on KPCA-ICSA-SVR established in this study has high prediction accuracy. However, further discussions are needed in terms of data scale and other aspects.
Due to the constraints of the underground working environment, future research needs to incorporate more influencing factors for a comprehensive analysis. Gas emission is a complex, dynamic process influenced by multiple factors. Beyond the indicators considered in this study, additional potential factors, such as surrounding rock structure and the microscopic properties of coal, should be examined for a more thorough understanding and prediction of gas emission levels.
The mechanical properties and structural morphology of the surrounding rock, especially around the coal seam, significantly impact gas emissions. The strength, hardness, and fracture distribution of the surrounding rock determine the coal seam’s stability and permeability. Softer or more fractured surrounding rock provides pathways for gas to seep out, leading to increased gas emissions. Additionally, structural deformations within the surrounding rock, such as faults and folds, may cause localized pressure changes in the coal seam, further influencing gas release. Stable surrounding rock can effectively restrict gas diffusion, while unstable formations exacerbate gas emissions, thereby increasing safety risks in the mine.
The microscopic properties of coal, particularly its pore structure and gas adsorption characteristics, also play a crucial role in gas emissions. The porosity, pore distribution, and particle size of coal directly affect its gas storage capacity. Coal seams with well-developed pore structures can store more gas and release it rapidly under changing external conditions. Furthermore, areas with high gas adsorption capacity tend to absorb more gas molecules, which are then released at an accelerated rate during pressure changes. Other micro-level characteristics, such as coal type, composition, coalification degree, and moisture content, also significantly influence gas release behavior.
In conclusion, while the predictive model in this study has shown certain advantages, integrating more influencing indicators will further enhance the model’s accuracy and practicality, providing better decision support for gas emission prediction and disaster prevention.
7. Conclusions
This study conducts an in-depth analysis of data cleaning, feature extraction, and model optimization for gas emission prediction, proposing a high-precision prediction model based on KPCA-ICSA-SVR. Through data processing, dimensionality reduction, and algorithmic enhancements, this approach effectively captures the nonlinear relationships within complex data, significantly improving prediction accuracy and the model’s generalization capability. The application of this model provides essential technical support for mine gas management.
(1) Data Cleaning and Feature Extraction: In this study, the Boxplot–MI method was employed for comprehensive data cleaning of gas emission data. After removing outliers with the Boxplot method, multiple imputation (MI) was used to handle missing data, ensuring data completeness and consistency. Next, Kernel Principal Component Analysis (KPCA) was used to extract key features and perform effective dimensionality reduction. The first four principal components, which explain a cumulative variance of 0.89016, were selected. This approach not only significantly simplified the data structure but also effectively preserved the essential information between variables.
(2) Model Optimization: An Improved Crow Search Algorithm (ICSA) was developed, integrating adaptive neighborhood search with reinforcement learning to optimize the SVR model’s hyperparameters. By adaptively adjusting the neighborhood range and employing reinforcement learning strategies for dynamic exploration of the global optimum, the ICSA efficiently and stably identified the optimal parameter combination for SVR, enhancing model convergence speed and accuracy.
(3) Comparative Experiment: A series of comparative experiments, including BP, KPCA-ICSA-BP, and conventional SVR models, were conducted to assess each model’s performance in gas emission prediction. Results demonstrated that the RMSE of the KPCA-ICSA-SVR model for the test set is 0.3132, which is lower than that of other models. Additionally, all other evaluation metrics show optimal performance, particularly in terms of prediction accuracy and generalization ability, surpassing the performance of other models.
Author Contributions
Conceptualization, L.L.; methodology, L.L. and X.M.; software, L.D.; validation, Y.C. and Y.J.; formal analysis, L.D. and X.M.; investigation, Y.J.; writing—original draft preparation, L.L. and L.D. All authors have read and agreed to the published version of the manuscript.
Funding
This research was funded by the Natural Science Foundation of Chongqing, grant number CSTB2022NSCQ-MSX1080; the Key Project of Science and Technology Innovation and Entrepreneurship Fund of Tiandi Technology Co., Ltd., grant number 2023-2-TD-ZD001; and the Science and Technology Project of Yulin, grant number CXY-2020-030.
Data Availability Statement
The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding authors.
Conflicts of Interest
The authors declare that this study received funding from the Key Project of Science and Technology Innovation and Entrepreneurship Fund of Tiandi Technology Co., Ltd. The funder was not involved in the study design, collection, analysis, interpretation of data, the writing of this article or the decision to submit it for publication.
References
- Wu, Y.; Wang, P.; Wang, D.; Bu, L.; Tian, J.; Sun, G.; Xu, S. Synergistic ductility deformation and helical design of carbon nanotube fiber composites. Carbon 2024, 229, 119441. [Google Scholar] [CrossRef]
- Wang, L.; Sun, Y.; Zheng, S.; Shu, L.; Zhang, X. How efficient coal mine methane control can benefit carbon-neutral target: Evidence from China. J. Clean. Prod. 2023, 424, 138895. [Google Scholar] [CrossRef]
- Harpalani, S.; Chen, G. Estimation of changes in fracture porosity of coal with gas emission. Fuel 1995, 74, 1491–1498. [Google Scholar] [CrossRef]
- Li, J.; Xu, T.; Lu, W.; Zhao, X.; He, Q.; Li, J.; Zhuo, H. Water-gas masking effect of the primary active sites in coal and room temperature oxidation of coal after desorption. J. China Coal Soc. 2024, 49, 2298–2314. [Google Scholar]
- Wang, K.; Li, K.; Du, F.; Zhang, X.; Wang, Y.; Sun, J. Research on prediction model of coal spontaneous combustion temperature based on SSA-CNN. Energy 2024, 290, 130158. [Google Scholar] [CrossRef]
- Song, J.; Deng, J.; Zhao, J.; Zhang, Y.; Wang, C.; Shu, C.M. Critical particle size analysis of gas emission under high-temperature oxidation of weathered coal. Energy 2021, 214, 118995. [Google Scholar] [CrossRef]
- Wu, Y.; Wang, P.; Wang, D.; Bu, L.; Tian, J.; Xu, S. Unveiling the microstructural evolution and interaction mechanisms for twisted structures. Int. J. Mech. Sci. 2024, 279, 109514. [Google Scholar] [CrossRef]
- Xie, J. Research on pre-pumping gas control technology for large-diameter borehole. J. Min. Sci. Technol. 2024, 9, 529–537+652. [Google Scholar]
- Guo, X.; Ren, Z.; Wang, Q.; Cui, C. Prediction of Gas Emission by BP Neural Network Based on Wavelet Analysis. IOP Conf. Ser. Earth Environ. Sci. 2019, 252, 052046. [Google Scholar] [CrossRef]
- Zhang, Z.; Ma, W. Prediction of gas emission in mining face based on random forest regression algorithm. J. Mine Autom. 2023, 49, 33–39. [Google Scholar]
- Yan, H. The Study on Prediction Method of Gas Emission Amount AQPSO-RBF in Fully Mechanized Mining Face and Its Application. Master’s Thesis, Xi’an University of Science and Technology, Xi’an, China, 2020. [Google Scholar]
- Zou, Q.; Chen, Z.; Cheng, Z.; Liang, Y.; Xu, W.; Wen, P.; Zhang, B.; Liu, H.; Kong, F. Evaluation and intelligent deployment of coal and coalbed methane coupling coordinated exploitation based on Bayesian network and cuckoo search. Int. J. Min. Sci. Technol. 2022, 32, 1315–1328. [Google Scholar] [CrossRef]
- Zhou, J.; Lin, H.; Jin, H.; Li, S.; Yan, Z.; Huang, S. Cooperative prediction method of gas emission from mining face based on feature selection and machine learning. Int. J. Coal Sci. Technol. 2022, 9, 51. [Google Scholar] [CrossRef]
- Zhao, W.; Dong, H.; Ren, J.; Yuan, Y.; Wang, K.; Wang, F. A software for calculating coal mine gas emission quantity based on the different-source forecast method. Int. J. Coal Sci. Technol. 2024, 11, 51. [Google Scholar] [CrossRef]
- Zhang, B.; Wang, H.; Wang, P.; Yu, G.; Gu, S. Experimental and theoretical study on the dynamic effective stress of loaded gassy coal during gas release. Int. J. Min. Sci. Technol. 2023, 33, 339–349. [Google Scholar] [CrossRef]
- Ma, L.; Qian, J.; Zhao, W.; Curtis, Z.; Zhang, R. Hydrogeochemical analysis of multiple aquifers in a coal mine based on nonlinear PCA and GIS. Environ. Earth Sci. 2016, 75, 716. [Google Scholar] [CrossRef]
- Saikia, J.; Saikia, P.; Boruah, R.; Saikia, B.K. Ambient air quality and emission characteristics in and around a non-recovery type coke oven using high sulphur coal. Sci. Total Environ. 2015, 530, 304–313. [Google Scholar] [CrossRef]
- Wang, Y.; Fu, H.; Zhang, Y. The identification model of coal and gas outburst intensity based on KPCA and CIPSO-PNN. Chin. J. Sens. Actuators 2015, 28, 271–277. [Google Scholar]
- Liu, S.; Qu, E.; Chun, L.V.; Zhang, X. Research on the prediction of blasting fragmentation in open-pit coal mines based on KPCA-BAS-BP. Sci. Rep. 2024, 14, 16804. [Google Scholar] [CrossRef]
- Guo, Q.; Chen, H.; Luo, J.; Wang, X.; Wang, L.; Lv, X.; Wang, L. Parameter inversion of probability integral method based on improved crow search algorithm. Arab. J. Geosci. 2022, 15, 180. [Google Scholar] [CrossRef]
- Hichem, H.; Rafik, M.; Ouahiba, C. New Discrete Crow Search Algorithm for Class Association Rule Mining. Int. J. Swarm Intell. Res. 2022, 13, 21. [Google Scholar] [CrossRef]
- Li, M.; Cheng, X. Identification and Prediction of Thermodynamic Disasters During Deep Coal Mining. Int. J. Heat Technol. 2022, 40, 1447. [Google Scholar] [CrossRef]
- Ledmi, M.; Ledmi, A.; Souidi, M.; Cherif, A.; Maarouk, T.; Cherif, C. High-utility itemsets mining integrating an improved crow search algorithm and particle search optimization. Soft Comput. 2024, 28, 8471–8496. [Google Scholar] [CrossRef]
- Li, K.; Wu, Y.; Wang, K.; Du, F.; Zhang, X.; Wang, Y.; Sun, J. Research on prediction of coal-gas compound dynamic disaster based on ICSA-CNN. Fuel 2024, 359, 130462. [Google Scholar] [CrossRef]
- Meng, Q.; Ma, X.; Zhou, Y. Forecasting of coal seam gas content by using support vector regression based on particle swarm optimization. J. Nat. Gas Sci. Eng. 2014, 21, 71–78. [Google Scholar] [CrossRef]
- Li, S.; Xu, K.; Xue, G.; Liu, J.; Xu, Z. Prediction of coal spontaneous combustion temperature based on improved grey wolf optimizer algorithm and support vector regression. Fuel 2022, 324, 124670. [Google Scholar] [CrossRef]
- Boula, A.; Christine, L.M.; Gunkel-Grillon, P.; Bour, O.; Selmaoui-Folcher, N. Potential contamination of stream waters by ultramafic mining sediments: Identification of geochemical makers (New Caledonia). J. Geochem. Explor. 2021, 232, 106879. [Google Scholar] [CrossRef]
- Farhangfar, A.; Kurgan, L.; Dy, J. Impact of imputation of missing values on classification error for discrete data. Pattern Recognit. 2008, 41, 3692–3705. [Google Scholar] [CrossRef]
- Zhu, Z. Prediction of Coal and Gas Outburst Based on PCA-BP Neural Network. China Saf. Sci. J. 2013, 23, 45–50. [Google Scholar]
- Zhang, F. Gas Emission Prediction Based on BP Neural Network. Informatiz. Res. 2016. [Google Scholar]
- Lan, T.; Guo, X.; Zhang, Z.; Liu, M. Prediction of microseismic events in rock burst mines based on MEA-BP neural network. Sci. Rep. 2023, 13, 9523. [Google Scholar] [CrossRef]
| Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).