Next Article in Journal
Low-Temperature Stress-Induced Changes in Cucumber Plants—A Near-Infrared Spectroscopy and Aquaphotomics Approach for Investigation
Previous Article in Journal
Comparative Study of Distributed Acoustic Sensing Responses in Telecommunication Optical Cables
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Intelligent Recognition of Muffled Blasting Sounds and Lithology Prediction in Coal Mines Based on RDGNet

School of Mining Engineering, Taiyuan University of Technology, Taiyuan 030024, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(24), 7601; https://doi.org/10.3390/s25247601
Submission received: 31 October 2025 / Revised: 28 November 2025 / Accepted: 11 December 2025 / Published: 15 December 2025
(This article belongs to the Section Intelligent Sensors)

Abstract

In the Yangquan coal mining region, China, muffled blasting sounds commonly occur in mine surrounding rocks resulting from instantaneous energy release following the elastic deformation of overlying brittle rock layers; they are related to fracture development. Although these events rarely cause immediate hazards, their acoustic signatures contain critical information about cumulative rock damage. Currently, conventional monitoring of muffled blasting sounds and surrounding rock stability relies on microseismic systems and on-site sampling techniques. However, these methods exhibit low identification efficiency for muffled blasting events, poor real-time performance, and strong subjectivity arising from manual signal interpretation and empirical threshold setting. This article proposes retentive depthwise gated network (RDGNet). By combining retentive network sequence modeling, depthwise separable convolution, and a gated fusion mechanism, RDGNet enables multimodal feature extraction and the fusion of acoustic emission sequences and audio Mel spectrograms, supporting real-time muffled blasting sound recognition and lithology classification. Results confirm model robustness under noisy and multisource mixed-signal conditions (overall accuracy: 92.12%, area under the curve: 0.985, and Macro F1: 0.931). This work provides an efficient approach for intelligent monitoring of coal mine rock stability and can be extended to safety assessments in underground engineering, advancing the mining industry toward preventive management.

1. Introduction

As the mining depth has increased, the stability of rocks surrounding deep roadways has emerged as a critical issue. In the Yangquan mining district of Shanxi Province, China, mining-induced low-frequency muffled blasting sounds are frequently emitted from within the coal-rock mass and are widely regarded as manifestations of stress concentration and microfracture propagation [1,2]. Detailed field monitoring at Yangsheng Coal Mine revealed that this deep site exhibits high in situ stress. Hence, these muffled blasting phenomena result from the abrupt release of accumulated elastic strain energy during the deformation and failure of the surrounding rock. This sudden energy release represents a common physical mechanism driving a continuum of mining-induced dynamic disturbances, from the muffled blasting sounds documented in the present study to severe rockbursts widely reported in deep coal mines worldwide [3,4,5]. Given that rockbursts, roof falls, and related dynamic disasters—many of which are preceded by similar acoustic precursors—remain a leading cause of fatalities and production losses in deep mining, the effective monitoring of such low-frequency acoustic signals is critical for early warning and safe production.
Surrounding rock stability is primarily analyzed based on manual inspections and using on-site sampling techniques, such as borehole coring and visual fracture assessment. However, these approaches suffer from low detection efficiency, poor timeliness, and high subjectivity. Under complex geological conditions, such limitations often lead to delayed or inaccurate hazard warnings. To overcome these shortcomings, Liu et al. [6] successfully identified early signals of stress variation in surrounding rocks using acoustic emission (AE) monitoring. Similarly, Wu et al. [7] highlighted the effectiveness of multisource data integration by combining microseismic monitoring with numerical simulations. Ye et al. [8,9] further developed a monitoring system for surrounding rock deformation based on omnidirectional structured-light vision sensors, significantly improving monitoring accuracy and automation.
In recent years, deep learning has rapidly outperformed traditional microseismic monitoring methods in rockburst prediction, achieving accuracy improvements of 15–30% on real mine datasets and markedly enhancing nonlinear modeling and multimodal data processing capabilities [10,11,12]. These advancements enable the development of more precise and intelligent disaster early-warning systems. In 2022, Pu et al. [13] proposed a rock burst intensity prediction model based on variable weights and matter–element extension theory, demonstrating its effectiveness in handling nonlinear rock mechanics parameters. In 2023, Li et al. [14,15] developed a multimodal data-driven deep learning model that integrated a convolutional neural network (CNN) with a long short-term memory model to process microseismic and stress data. Their approach improved model robustness but exhibited suboptimal accuracy when applied to small datasets. In 2021, a real-time rock burst intensity prediction model was introduced, combining a CNN, the Adam optimization algorithm, and Bayesian optimization, further enhancing prediction efficiency and convergence performance. In 2025, Li et al. [16] highlighted the advantages of intelligent machine learning for rock burst prediction, emphasizing that enhanced high-dimensional data processing capabilities significantly improve diagnostic performance. Wang et al. 2023 [17] addressed geological uncertainties via a stacking integration ensemble of multimodal predictors (e.g., seismic, stress, and lithological features), achieving a classification accuracy of >90%. Basnet et al. [18,19] applied ensemble learning techniques to mitigate model overfitting. In 2024, Liu et al. [20] proposed the NGO–CNN–BiGRU–Attention model, which enhanced feature extraction and temporal representation through attention mechanisms. In 2025, Zhang et al. [21] introduced a prototype-based algorithm for rock burst type identification and risk prediction, substantially improving classification robustness. Collectively, these studies demonstrate that deep learning has become the dominant approach in rock burst prediction, offering reliable and efficient support for mine safety management.
Deep learning has been increasingly applied to process multisource sensor data for automated rock classification [22,23,24,25,26,27]. In 2021, exploration efficiency was improved by applying a CNN to rock image analysis [28,29,30,31,32]. In 2022, employed ResNet34 for rock image classification, optimizing feature extraction through data augmentation and residual learning, and achieved an accuracy of 88.1% [33,34,35]. Between 2023 and 2024, developed a CNN architecture that attained 91.03% accuracy in three-class rock identification and enhanced model robustness through cross-validation [36,37,38,39,40,41]. These advances, including recent multimodal approaches [42], have significantly improved classification accuracy, reduced manual dependency, and accelerated the intelligent transformation of the mining industry.
In summary, numerous studies have achieved significant progress in rock burst prediction and lithology identification through deep learning techniques. However, challenges remain in effectively processing and fusing multisource signals. To address these limitations, this study proposes a multimodal deep learning model based on the retentive depthwise gated network (RDGNet), utilizing AE signals and acoustic signals as input data sources. The proposed model performs multiscale feature extraction and dynamic data fusion of AE signals and audio Mel spectrograms by integrating retentive network (RetNet) sequence modeling, depthwise separable convolution, and gated fusion mechanisms.
The main contributions of this study are summarized as follows:
For the first time, RetNet sequence modeling is combined with depthwise separable convolution to process multimodal signals—AE sequences and Mel spectrograms—of muffled blasting sounds from roadway surrounding rocks. This approach enables end-to-end feature extraction and classification, thereby improving both identification accuracy and real-time performance.
A RetNet variant incorporating rotary position embedding (RoPE) and group normalization (GN) is introduced to strengthen temporal feature representation. In addition, a gated fusion mechanism employing Sigmoid-weighted gates dynamically optimizes feature fusion, enhancing model robustness under complex geological conditions.
The synergistic contributions of each module are verified through ablation experiments on the sequence, convolutional, and fusion branches. The proposed RDGNet achieves 92.12% accuracy, an area under the curve (AUC) of 0.985, and a Macro F1 score of 0.931 on the Yangsheng Coal Industry dataset, outperforming baseline models, particularly under low signal-to-noise ratio (SNR) conditions.
Multiclass mixed-input experiments demonstrate the stability of RDGNet and its low sensitivity to lithology distribution, with classification accuracies exceeding 90% across all lithology types. These results indicate that the model is not only effective for monitoring surrounding rocks in coal mines but is also adaptable to other underground engineering applications, such as tunnel stability assessment.

2. Theoretical Background

2.1. Depthwise Separable Convolution

Depthwise separable convolution is an efficient convolutional operation designed to reduce the parameter count and computational complexity of traditional CNNs, as illustrated in Figure 1. This operation decomposes a standard convolution into two sequential stages: depthwise convolution and pointwise convolution. In the depthwise convolution stage, individual convolution kernels are applied independently to each input channel to extract intrachannel spatial features. In the subsequent pointwise convolution stage, a 1 × 1 convolution is used to fuse the multichannel outputs, capturing interchannel correlations. This decomposition markedly decreases computational cost and memory usage while maintaining strong representational capacity. In this study, depthwise separable convolution is employed in the convolutional branch that processes audio Mel spectrograms, enabling efficient multimodal feature extraction and enhancing the model’s computational efficiency and real-time performance.
The computational complexity of a standard convolution can be expressed as follows:
C s t d = H × W × N × K 2   ,
where H and W denote the height and width of the input feature map, respectively; N represents the number of output channels; and K is the kernel size.
Similarly, the computational complexity of a depthwise separable convolution is given by
C s e p = H × W × M × K 2 + H × W × M × N     ,
where M represents the number of input channels.

2.2. RetNet Module

RetNet is a novel sequence modeling architecture designed to efficiently capture long-range dependencies while maintaining low computational cost. As illustrated in Figure 2, the core of RetNet lies in its retention mechanism, which unifies three computational paradigms: parallel representation for efficient training, recurrent representation for low-cost inference, and chunkwise recurrent representation for long-sequence processing. Through its multiscale retention mechanism, RetNet effectively models contextual dependencies across different temporal scales, thereby avoiding the quadratic complexity characteristic of Transformer attention mechanisms. Specifically, the retention mechanism incorporates RoPE and gating operations to model temporal dependencies while achieving O(1) space complexity during inference. In this study, RetNet serves as the multimodal backbone of the proposed framework. The sequence branch leverages RetNet to extract temporal features from AE signals, enabling efficient multimodal fusion and improving both recognition accuracy and real-time performance in muffled blasting sound identification.
RetNet employs a retention mechanism to model sequential data, enabling parallelized training and recurrent inference. The mathematical formulation of the retention mechanism is expressed as follows:
o t =   q t   · S t ,
S t = γ S t 1 + k t v t T   ,
where γ denotes the decay factor, v t is the velocity component at time step t , and T refers to the discrete time index. This mechanism effectively captures temporal dependencies in AE signals.

2.3. Gated Fusion

Gated fusion is a multimodal information fusion strategy that dynamically adjusts the weights of different modality features through a gating mechanism, thereby achieving adaptive and robust fusion. Its framework is illustrated in Figure 3. The gating mechanism—typically implemented using a Sigmoid activation function—generates weighted gates based on the information entropy or correlation of the input modalities. This enables adaptive feature integration while mitigating the influence of noisy or redundant modalities. In this study, gated fusion is applied to the fusion layer of RetNet to dynamically integrate AE sequence features and audio convolutional features, thereby enhancing the model’s robustness and classification accuracy for multisource signals.
Gated fusion dynamically integrates modal features using the following formula:
g =   σ W g · f 1 : f 2   ,
where σ denotes the Sigmoid function, W g represents the learnable weight matrix for the gate, and f i i = 1 ,   2 denotes the feature vector from the i-th branch.
The fused feature is given by
f f u s e d = g · f 1   + 1 g · f 2 .
This mechanism facilitates the adaptive fusion of AE and audio features, improving the overall robustness of the proposed model.

3. Materials and Methods

3.1. RDGNet Multimodal Model Design

RDGNet is an efficient multimodal sequence modeling architecture that supports parallel training and low-cost inference. It is specifically designed to process multimodal signals generated by muffled blasting sounds in roadway surrounding rocks. The model integrates sequence branches, convolution branches, and gated fusion layers to enable robust extraction and fusion of AE signals and audio features. The overall architecture of RDGNet is illustrated in Figure 4.
The sequence branch first processes AE input signals through a local enhancement layer, which applies one-dimensional convolution k e r n e l _ s i z e = 3 ,   p a d d i n g = 1 to extract local temporal features:
x = R e L U B N C o n v 1 D x ,
where x denotes the AE input sequence (batch × window samples × 3), containing three-dimensional features: amplitude, peak frequency, and energy. Subsequently, lithology embeddings—encoded using one-hot encoding and projected to a 256-dimensional space—are incorporated and passed into the RetNet variant layers.
The RetNet variant adopts RoPE to enhance positional sensitivity, defined as
q = q · cos θ + q · sin θ · R   ,
where θ is the frequency vector and R denotes the rotation matrix. This variant supports a multihead retention mechanism (num_heads = 4), selected via hyperparameter tuning. The similarity matrix in parallel computation mode is expressed as
s i m = q k T · d e c a y m a s k     .
To capture long-term dependencies, the decay mask employs logarithmically distributed decay factors γ (−3.0 to −0.1). After three stacked RetNet variant layers, the output sequence features are flattened and projected to 128 dimensions using a linear transformation layer.
The convolutional branch processes Mel spectrograms by first applying depthwise separable convolution to extract spatial features. The first convolutional layer ( k e r n e l _ s i z e = 3 ) is computed as
f e a t 1 = P o i n t w i s e D e p t h w i s e m e l   .
This stage is followed by adaptive max pooling and a second convolutional layer ( k e r n e l _ s i z e = 5 ) , which incorporates residual connection:
f u s e d   c o n v = C o n v f e a t 1 + f e a t 1   .
The resulting feature maps are projected to 128 dimensions using a 1 × 1 convolution, followed by batch normalization, rectified linear unit activation, and pooling.
The gated fusion layer dynamically integrates the sequence features f s and convolutional features f c through a learnable gate, defined as
g = σ   W g · f s : f c .
The fused feature is given by
f f u s e d = g · f s + 1 g · f c   .
Finally, classification probabilities are generated via a fully connected (linear) layer. The model is optimized using the Focal Loss function, with focusing parameters γ = 2.0 and α = 0.5 Additionally, lithology-weighted sampling is employed to ensure balanced representation of minority lithology classes during training.

3.2. RetNet Variant

The structure of the RetNet variant is illustrated in Figure 4. The input sequence x is first linearly projected to obtain the query (Q), key (K), and value (V) representations. These are subsequently processed through the retention mechanism, which operates in either parallel or recurrent mode, followed by GN and projection layers.
Compared with the standard RetNet, this variant preserves the core masked self-attention and exponential decay mechanisms but introduces several key improvements: RoPE to enhance positional encoding, GN for stabilized gradient propagation, head-specific initialization for improved convergence, and explicit mode switching for flexible computation. Collectively, these enhancements improve positional sensitivity, training stability, and computational efficiency.
The RetNet variant integrated into the RDGNet multimodal model differs from the standard RetNet by incorporating adaptive modifications tailored to the specific requirements of multimodal muffled blasting signal analysis.
The standard RetNet employs complex-valued formulations to implicitly encode relative positional information and utilizes an exponential decay matrix D to preserve temporal dependencies. However, this approach can exhibit limited sensitivity to positional shifts when processing irregular or noise-dense data, such as AE sequences, particularly under data augmentation conditions. In the variant RetNet, RoPE is explicitly introduced between the generation of Q and K vectors and the subsequent similarity computation, as illustrated in the figure. The qkv projection block branches into Q, K, and V, with Q and K entering the “QK” block, where RoPE is applied. Specifically, the Q and K tensors are reshaped to the form (B, N, D) and processed using a custom rotary embedding function defined as
R o P E T = T : d / 2 cos θ T d   / 2 : sin θ T : d / 2 sin θ + T d   / 2 : cos θ
where d denotes the head dimension, θ represents the predefined frequency vector ( θ k = t · 10,000 2 k / d   ,   k = 0 ,   1   · · · d 2 1 , t   f o r   t h e   p o s i t i o n   i n d e x ) . This operation enforces rotational invariance within the dot product space, thereby complementing the decay-based retention mechanism of the standard RetNet and enhancing the model’s capability to capture fine-grained temporal dynamics.
Compared with the standard RetNet, the proposed variant preserves the core masked self-attention and exponential decay mechanisms but integrates several key enhancements—RoPE, GN, head-specific initialization, and explicit mode switching—to improve positional awareness, training stability, and computational efficiency. While the standard RetNet employs GN (instead of layer normalization) to stabilize activation distributions, the variant further refines this process by applying per-head normalization to the reshaped outputs, ensuring independent standardization within each attention head. This design effectively mitigates the influence of irregular noise in multimodal signals, resulting in more stable convergence during training.
In the standard RetNet, multiscale decay factors ( γ ) are assigned to each head to capture temporal dependencies at different scales. In contrast, the variant allocates unique decay factors that transition linearly from strong to weak decay, enabling parallel modeling of multiscale temporal dynamics. This configuration is particularly advantageous for fusing temporal AE sequences with Mel spectrogram features, as it enhances cross-modal coherence and improves the F1 score by approximately 4% in class-imbalanced tasks, owing to its gradient design that increases sensitivity to minority-class signals.
Furthermore, like the standard RetNet, the variant retains support for parallel training and recurrent inference. However, in the standard RetNet, mode switching is typically implemented implicitly within the computational formulas, which can introduce efficiency losses in mixed training–inference pipelines. In contrast, the variant adopts an explicit mode-switching mechanism controlled by dedicated parameters. For example, it utilizes Einstein summation (einsum) operations for efficient computation, transitioning from “SM” through “D” to “Sim” in parallel mode, while iteratively updating states from the Q, K, and V branches in recurrent mode. The framework defaults to parallel mode during training to maximize throughput and switches to recurrent mode during inference to reduce memory consumption. This explicit control mechanism significantly decreases inference latency while preserving compatibility with the original RetNet’s computational complexity. Collectively, these optimizations enhance the model’s robustness, real-time performance, and stability when processing noise-dense AE sequences in multimodal fusion tasks.

4. Results

4.1. Data Collection and Preprocessing

This study utilized muffled blasting sound data from roadway surrounding rocks in the Yangsheng Coal Industry as the engineering background. The coal seam exhibited a synclinal structure, with higher elevations to the east and west and a relatively lower central region, characterized by undulating seam fluctuations. The average seam thickness was 5.72 m, with a density of 1.43 t/m3, and contained zero to four gangue interlayers with an average dip angle of 5°. The immediate roof comprised mudstone, sandy mudstone, and, locally, siltstone or fine sandstone, while the old roof consisted mainly of limestone with well-developed joints and fractures. The floor was composed of carbonaceous mudstone, locally transitioning to sandy or aluminous mudstone. As shown in the borehole lithology histogram (Figure 5), the top of the coal seam predominantly consisted of limestone, mudstone, and sandstone layers. Accordingly, 20 limestone, 15 sandstone, and 10 mudstone samples were collected as the core of the laboratory dataset. To account for geological heterogeneity and improve the model’s generalization capability, five shale samples were additionally included in the experimental group.
An eight-channel integrated AE detection system was employed to record event parameters and positional information during the uniaxial compression tests of rock specimens. The recording equipment continuously captured acoustic signals under laboratory conditions designed to simulate in situ stress environments. Each lithology specimen exhibited distinct brittle failure characteristics under uniaxial loading, as illustrated in Figure 6. During compression, intermittent AE signals were generated as microcracks initiated and propagated, followed by intense sound bursts corresponding to the onset of macrofracture and ultimate failure. The AE data collected by the integrated detector were stored as temporal sequences, including timestamps, amplitudes, peak frequencies, and energy values. Meanwhile, the audio signals were recorded as continuous waveform data and subsequently processed using the Librosa Python (Version: 0.11.0) library for feature extraction and Mel spectrogram generation.
The relationships among audio signals, stress, AE energy, and time during the compression–rupture process are illustrated in Figure 7. The AE sampling rate was set to 1000 Hz, while the audio signals were recorded at their native sampling rate. Data preprocessing involved time synchronization, filtering and denoising, normalization, and signal segmentation using a 100 ms time window. To mitigate class imbalance and enhance generalization, data augmentation techniques were applied, including time stretching, noise injection, signal flipping, and missing-data simulation. Specifically, Gaussian noise was introduced with normalized standard deviations of 0.03 for AE signals and 0.003 for Mel spectrograms, while missing-data rates ranged from 0.10 to 0.15. These augmentations primarily targeted positive samples and minority lithologies (e.g., shale) to balance the dataset distribution. The details of the sampling distribution across lithologies before and after data augmentation as well as the final composition of the training, validation, and test sets are provided in Table 1. To ensure rigorous evaluation and eliminate any risk of data leakage, the dataset was first stratified and divided at the specimen level into training (70%), validation (15%), and test (15%) sets. All data augmentation operations were subsequently performed exclusively for the training set. The data augmentation techniques employed in this study are summarized in Table 2, along with their detailed descriptions, parameters, and target modalities.
The processed AE signals were fed into the sequence branch of RDGNet, whereas the audio signals were converted into Mel spectrograms using a short-time Fourier transform with parameters n_mels = 128, hop_length = 512, and n_fft = 2048, as illustrated in Figure 8. These spectrograms were then input into the convolutional branch. The RDGNet model comprised 4,147,224 learnable parameters. Model training took approximately 840 s on a system equipped with a 12th Gen Intel Core i5-12600KF CPU and an NVIDIA GeForce RTX 3080 GPU. The network achieved multimodal feature fusion and supported end-to-end learning, enabling joint optimization of AE and audio modalities.

4.2. Feature Visualization Analysis

To intuitively analyze the influence of the gated fusion mechanism on the recognition of muffled blasting sounds from roadway surrounding rocks, rock specimens collected from the Yangsheng Coal Industry were selected as representative samples. The t-distributed stochastic neighbor embedding (t-SNE) algorithm was employed to visualize the feature representations learned at different stages of the RDGNet model. Specifically, the test data were first input into the trained RDGNet, and feature maps were extracted from multiple layers of the network. The t-SNE algorithm then projected these high-dimensional feature maps onto a two-dimensional space, where each feature was represented as a point in the coordinate system. The resulting scatter plots were generated based on these two-dimensional coordinates.
To evaluate the effectiveness of the gated fusion mechanism, three types of feature representations were selectively extracted: (1) features learned through data augmentation, (2) features obtained from the sequence branch, and (3) features optimized by the gated fusion mechanism. As illustrated in Figure 9, after feature extraction by the sequence branch, the limestone and sandstone samples were almost completely separated from other lithologies, whereas considerable overlap remained between the mudstone and shale samples. This observation suggests that the preliminary sequence features had limited discriminative ability for minority lithologies. After modal weight recalibration through the gated fusion mechanism, the number of overlapping samples was markedly reduced, demonstrating an improvement in lithology discrimination. The final classifier outputs, shown in Figure 9c, exhibit complete separation among the four lithologies, with distinct cluster boundaries and no overlap. These findings confirm that the proposed RDGNet framework effectively captures multimodal relationships in muffled blasting sounds and achieves accurate lithology classification.
Additionally, to assess the convergence and stability of the model during training, the variations in accuracy and loss across epochs were analyzed. As shown in Figure 10, the training accuracy increased steadily after approximately 50 epochs, eventually stabilizing around 92%, while the loss decreased sharply in the early training stages and subsequently converged. This trend indicates effective learning and minimal overfitting. Moreover, the visualization highlights the efficiency of the RDGNet architecture in processing multimodal data, with the RetNet variant and gated fusion mechanism jointly enhancing gradient flow and feature integration.

4.3. Ablation Study

To systematically evaluate the contribution of each component within the RDGNet architecture, several ablation experiments were conducted by sequentially removing or replacing the sequence branch, the convolution branch, the RetNet module, the Focal Loss function, the gated fusion mechanism, and data augmentation as well as introducing variants such as the standard RetNet, concatenation fusion, and average fusion models. The experiments were performed on the augmented dataset, trained for 30 epochs with a batch size of 512, using the AdamW optimizer with a learning rate of 5 × 10−5. Evaluation metrics included the area under the receiver operating characteristic (ROC) curve (AUC–ROC), Macro F1 score, accuracy, positive and negative recall, and the optimal decision threshold (determined via ROC curve optimization). The complete RDGNet configuration served as the baseline model.
As shown in Figure 11 and Table 3, the baseline RDGNet achieved the highest overall performance, with an AUC of 0.985, a Macro F1 score of 0.931, and an accuracy of 0.93. At an optimal threshold of 0.7, the positive and negative recalls were 0.93 and 0.932, respectively, indicating effective class balance without overfitting to positive samples. When the sequence branch was removed, the AUC decreased to 0.9334, the Macro F1 dropped to 0.8959, and the positive recall fell to 0.89, while the threshold increased to 0.705. This degradation underscores the RetNet module’s critical role in capturing temporal dependencies from AE sequences. The observed performance differences were statistically significant according to a t-test (p < 0.05), confirming the reliability and necessity of each RDGNet component.
Furthermore, to validate improvements achieved by the RetNet variant (which involved RoPE, GN, and head-specific initialization), we conducted a direct comparison with the standard RetNet over five independent runs. As shown in Table 4, the variant outperformed the standard RetNet, achieving a higher AUC, Macro F1 score. Additionally, the variant demonstrated higher efficiency than the standard RetNet, with an average time per epoch of 14.0 s compared with 16.8 s for the standard RetNet. These results substantiate the variant’s contributions to enhanced temporal dependency modeling—improving AUC by ~2.11%, Macro F1 by 2.45%, and accuracy by 1.09%—and to computational efficiency by reducing the average time per epoch and total training time by ~16.67%.

4.4. Multiclass Data Input Experiments

To realistically simulate the complex field scenario in which multiple lithologies coexist and damage can occur asynchronously, a dedicated multilithology mixed-signal experiment was performed.
Ten thousand mixed samples were synthetically generated by linearly superimposing 1–4 lithologies (random selection without replacement; proportions of samples with 1, 2, 3, or 4 lithologies: 20%, 30%, 30%, and 20%, respectively) using Dirichlet-distributed weights α = 1.0 . Each mixed sample was labeled with a genuine four-dimensional, multilabel damage vector y = [ y 1 ,   y 2 ,   y 3   , y 4 ] T ϵ 0,1 } 4 , where y i = 1 if the corresponding lithology component exhibited muffled blasting or macroscopic shear failure in its original record.
For this specific experiment only, the original RDGNet was minimally modified by replacing the single binary muffled blasting head with four independent lithology-specific damage heads (one Sigmoid unit per rock type: limestone, sandstone, mudstone, and shale). The pre-extracted rise time/amplitude and average frequency values—classical indicators of tensile vs. shear cracking in rock mechanics—were retained in the auxiliary branch and proved particularly valuable for distinguishing between failure modes across superimposed lithologies. All other network components (such as RetNet sequence branch, convolutional Mel branch, gated fusion) remained unchanged.
Training was conducted using per-dimension binary cross-entropy with batch-balanced Focal Loss; the four damage heads shared a total loss weight of 6.0. On a hold-out test set of 2000 unseen mixtures, the adapted model achieved an exact match ratio of 82.3%, a Macro F1 score of 0.938, and a Hamming loss of 0.071. The per-lithology F1 scores ranged from 0.912 (limestone) to 0.951 (mudstone). The performance remained robust (>79% exact match), even when four lithologies were superimposed simultaneously.
The comprehensive performance and interpretability of the proposed RDGNet are illustrated in Figure 12, Figure 13, Figure 14 and Figure 15. Figure 12 shows the confusion matrices for the damage prediction of each lithology; Figure 13 reports the F1 score, AUC, precision, and recall across the four rock types; Figure 14 displays three representative mixed-lithology seismic waveforms with corresponding RDGNet multilabel damage predictions; and Figure 15 shows the learned multisource fusion gate weights for the first 30 test samples.
These results (Table 5 and Figure 12, Figure 13, Figure 14 and Figure 15) demonstrate that with only a minor and physically meaningful extension of the output layer, the original RDGNet architecture can effectively disentangle contributions from multiple overlapping lithologies and independently predict the damage state of each rock type, directly addressing practical monitoring needs in layered surrounding rocks.

4.5. Noise Experiments and Model Comparison

To evaluate the robustness of the RDGNet model under noisy conditions and its superiority over baseline architectures, this section presents two experimental analyses: noise experiments and model comparisons. In the noise experiments, Gaussian noise was added to the test set at SNR levels of −20, −10, 0, 20, and 40 dB, and performance was evaluated using AUC and Macro F1 metrics. For the model comparison, several baseline models were implemented: TransformerNet (replacing the RetNet module with a Transformer layer), MambaNet (replacing RetNet with a Mamba layer), TransformerXLNet (replacing RetNet with a TransformerXLNet layer), and FlashTransformerNet (adopting an FlashAttention + Transformer structure in the sequence branch and a multiscale convolution module in the convolution branch). These baselines were chosen as representative sequence modeling alternatives to demonstrate the advantages of RetNet-based sequence processing. All models were trained using the same augmented dataset and identical hyperparameter settings (30 epochs, batch size = 512, learning rate = 5 × 10−5, AdamW optimizer).
The noise experiment results demonstrated that RDGNet maintained high performance across varying noise levels. At a high SNR of 40 dB, the model achieved an AUC of 0.9869 and a Macro F1 score of 0.9356. Even at a low SNR of −20 dB, it sustained an AUC of 0.9805 and a Macro F1 of 0.9148, exhibiting only minor performance degradation and outperforming all baseline models. In contrast, FlashTransformerNet experienced the most significant deterioration at low SNR (AUC = 0.970, F1 = 0.902), while TransformerNet and MambaNet showed larger declines (F1 = 0.908 and 0.911, respectively). TransformerXLNet performed relatively better (F1 = 0.917) but still fell short of RDGNet. The inclusion of data augmentation further mitigated noise-induced effects, confirming the robustness and generalization ability of RDGNet.
Comparison results with other models indicate that, in the model comparison experiments, RDGNet again achieved the highest overall performance (AUC = 0.987, Macro F1 = 0.938). The comparative results were as follows: TransformerNet—AUC = 0.984, F1 = 0.930; MambaNet—AUC = 0.982, F1 = 0.930; TransformerXLNet—AUC = 0.984, F1 = 0.930; and FlashTransformerNet—AUC = 0.9740, F1 = 0.9079. Furthermore, RDGNet exhibited the lowest floating-point operation count, demonstrating its computational efficiency and validating the superiority of its multimodal fusion architecture.
Detailed experimental outcomes are presented in Figure 16 and Figure 17.

5. Conclusions and Limitations

This study presents RDGNet, a deep learning model developed for the real-time identification of muffled blasting sounds and lithology classification in coal mine roadways. By integrating RetNet-based sequence modeling, depthwise separable convolution, and a gated fusion mechanism, RDGNet achieves robust multimodal feature extraction and fusion from AE sequences and Mel spectrograms. Validated on the Yangsheng Coal Mine dataset, RDGNet achieved an accuracy of 92.12%, an AUC of 0.985, and a Macro F1 score of 0.931, outperforming all baseline models. Even under low-SNR conditions (−20 dB), it maintained strong performance (F1 = 0.914), demonstrating high robustness to noise. The model effectively supports surrounding rock damage assessment and support parameter optimization, thereby contributing to enhanced mine safety and operational reliability.

5.1. Conclusions

The innovation of this study lies in the first application of a RetNet variant to multimodal monitoring of roadway surrounding rocks, integrating rock mechanics principles with advanced deep learning architectures to construct an efficient and robust network. The proposed model markedly enhances the accuracy and real-time performance of muffled blasting sound recognition, providing reliable data support for surrounding rock damage assessment and optimization of support parameters. Experimental findings demonstrate RDGNet’s clear advantages in feature extraction, classification accuracy, and robustness.
  • The sequence features extracted by RDGNet form distinct lithology clusters in low-dimensional space, confirming its strong ability to differentiate rock types and accurately capture subtle variations in muffled blasting sounds across lithologies.
  • The RetNet variant significantly enhances temporal dependency modeling, the gated fusion mechanism dynamically optimizes and integrates multimodal features, and data augmentation strategies combined with Focal Loss effectively mitigate class imbalance and noise interference, ensuring stable learning and minimal performance degradation.
  • Under laboratory-simulated operational conditions, RDGNet achieved an overall failure prediction accuracy of 92.12%, with balanced classification performance across lithologies. All rock types exhibited AUC values exceeding 0.951, and Macro F1 scores ranging from a maximum of 0.927 (shale) to a minimum of 0.883 (sandstone), demonstrating low sensitivity to variations in lithology distribution and stable performance in random mixed-signal environments.
  • At low-SNR conditions (−20 dB), RDGNet maintained an F1 score of 0.914 with minimal performance decline while achieving the lowest computational overhead, underscoring its superior adaptability and resource efficiency in noise-intensive geological environments.

5.2. Limitations

Although RDGNet performs well on the current dataset, there are two main limitations. First, the dataset comprises data collected from uniaxial compression tests of only 50 laboratory specimens (45 specimens with the in situ lithology shown in Figure 5 and 5 shale samples used for sample diversity). The failure mechanisms observed in the laboratory inevitably differ from those under complex in mine triaxial stress, creep, and groundwater conditions. Preliminary tests on real underground background noise from an operational coal mine showed a modest reduction in accuracy and Macro F1 score compared with the synthetic Gaussian noise results reported earlier.
To overcome these limitations, an identical-sensor long-term monitoring system has been deployed at an active mining face in the Yangsheng mining area. Over 5000 real-world labeled events will be recorded within the next 12 months for domain adaptation and continual learning by future RDGNet versions.

Author Contributions

G.L. and H.D. conceived, designed, and performed the experiments. K.W. and X.Z. analyzed the data and administrated the project. G.L. and J.S. implemented a part of the source code of the algorithm and wrote the paper. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (52474107).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are not publicly available due to privacy/ethical restrictions. However, the data are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AEAcoustic emission
AUCArea under the curve
CNNConvolutional neural network
GNGroup normalization
ROCReceiver operating characteristic
RoPERotary position embedding
SNRSignal-to-noise ratio

References

  1. Zhao, S.; Chao, Q.; Yang, L.; Qin, K.; Zuo, J. A review on application of acoustic emission in coal–analysis based on CiteSpace knowledge network. Processes 2022, 10, 2397. [Google Scholar] [CrossRef]
  2. Su, Z.; Jing, S.; Xie, W.; Tang, Q. Research on coal acoustic emission characteristics and damage evolution during cyclic loading. Front. Earth Sci. 2022, 10, 818452. [Google Scholar] [CrossRef]
  3. Zhang, L.; Liu, G.; Wei, X.; Zhang, Y. Mechanical properties and acoustic emission evolution of water-bearing sandstone under triaxial conditions. Front. Earth Sci. 2023, 11, 1212095. [Google Scholar] [CrossRef]
  4. Di, Y.; Wang, E.; Li, Z.; Liu, X.; Huang, T.; Yao, J. Comprehensive early warning method of microseismic, acoustic emission, and electromagnetic radiation signals of rock burst based on deep learning. Int. J. Rock Mech. Min. Sci. 2023, 170, 105519. [Google Scholar] [CrossRef]
  5. Zhang, M.; Fan, J.; Du, J.; Jiang, D.; Chen, J.; Yuan, Q.; Hao, L.; Wang, Y. Experimental study on effects of load damage precursor information and response characteristic of gas-containing coal for mining safety based on acoustic emission. Process Saf. Environ. Prot. 2024, 184, 993–1010. [Google Scholar] [CrossRef]
  6. Liu, H.; Xu, F.; Liu, B.; Deng, M. Time-series prediction method for rockburst hazard level based on CNN-LSTM. J. Cent. South Univ. 2021, 52, 659–670. [Google Scholar]
  7. Wu, H.; Li, Q.; Zhu, C.; Tang, P. Stability analysis of surrounding rock in mining tunnels based on microseismic monitoring and numerical simulation. Sustainability 2025, 17, 630. [Google Scholar] [CrossRef]
  8. Ye, Q.; Wang, B.; Jiang, W.; Deng, X.; Tao, H.; Liu, J.; He, W. Research on deformation monitoring method for surrounding rock in roadway based on an omnidirectional structured light vision sensor system. Measurement 2025, 235, 117867. [Google Scholar] [CrossRef]
  9. Liu, Y.; Gao, F.; Yan, W.; Duan, Z.; Chen, G.; Jiang, F.; Tian, Z.; Li, J. Study on acoustic emission evolution features of gas-bearing coal under different gas pressures. Sci. Rep. 2025, 15, 29120. [Google Scholar] [CrossRef]
  10. Chen, J.; Chen, Y.; Yang, S.; Zhong, X.; Han, X. A prediction model on rockburst intensity grade based on variable weight and matter-element extension. PLoS ONE 2019, 14, e0218525. [Google Scholar] [CrossRef]
  11. Afraei, S.; Shahriar, K.; Madani, S.H. Developing intelligent classification models for rock burst prediction: Comparison of performance among statistical and machine learning approaches. Eng. Geol. 2019, 255, 82–96. [Google Scholar]
  12. Zhao, H.; Chen, B. Data-driven model for rockburst prediction. Shock Vib. 2020, 2020, 5735496. [Google Scholar] [CrossRef]
  13. Pu, Y.; Apel, D.B.; Xu, H. Machine learning versus conventional techniques for rockburst risk assessment: A comprehensive review. Int. J. Min. Sci. Technol. 2022, 32, 921–936. [Google Scholar]
  14. Li, N.; Jimenez, R.; Zhu, H. Comparison of machine learning and traditional methods for rockburst prediction in deep hard rock mines. Rock Mech. Rock Eng. 2023, 56, 123–139. [Google Scholar]
  15. Li, J.; Fu, H.; Hu, K.; Chen, W. Data preprocessing and machine learning modeling for rockburst assessment. Sustainability 2023, 15, 13282. [Google Scholar] [CrossRef]
  16. Li, F.; Li, D.; Gu, Y. Research into acoustic emission monitoring in similar simulation of coal mining under roof-confined water drainage. Processes 2025, 13, 3287. [Google Scholar] [CrossRef]
  17. Wang, J.C.; Ma, H.J.; Yan, X.H. Rockburst intensity classification prediction based on multi-model ensemble integration stacking. Mathematics 2023, 11, 838. [Google Scholar] [CrossRef]
  18. Basnet, P.M.S.; Jin, A.; Mahtab, S. Applying machine learning approach in predicting short-term rockburst risks using microseismic information: A comparison of parametric and non-parametric models. Nat. Hazards 2025, 121, 731–758. [Google Scholar] [CrossRef]
  19. Basnet, P.M.S.; Mahtab, S.; Jin, A. A comprehensive review of intelligent machine learning based predicting methods in long-term and short-term rock burst prediction. Tunn. Undergr. Space Technol. 2023, 142, 105434. [Google Scholar] [CrossRef]
  20. Liu, H.Y.; Ma, T.; Lin, Y.; Peng, K.; Hu, X.; Xie, S.; Luo, K. Deep learning in rockburst intensity level prediction: Performance evaluation and comparison of the NGO-CNN-BiGRU-attention model. Appl. Sci. 2024, 14, 5719. [Google Scholar] [CrossRef]
  21. Zhang, X.; Li, G.; Chen, Y.; Wang, H.; Zhang, H.; Li, H.; Du, W.; Li, X.; Xu, X.; He, Y. A prototype-based rockburst types and risk prediction algorithm considering intra-class variance and inter-class distance of microseismic data. Front. Earth Sci. 2025, 13, 1601090. [Google Scholar] [CrossRef]
  22. Yin, X.; Liu, Q.; Huang, X.; Pan, Y. Real-time prediction of rockburst intensity using an integrated CNN-Adam-BO algorithm. Tunn. Undergr. Space Technol. 2021, 117, 104184. [Google Scholar] [CrossRef]
  23. Adoko, A.C.; Gokceoglu, C.; Wu, L.; Zuo, Q.J. Knowledge-based and data-driven fuzzy modeling for rockburst prediction. Int. J. Rock Mech. Min. Sci. 2013, 61, 86–95. [Google Scholar] [CrossRef]
  24. Zhang, Y.; Fang, K.; He, M.; Liu, D.; Wang, J.; Guo, Z. Rockburst prediction using artificial intelligence techniques: A review. Rock Mech. Bull. 2024, 3, 100129. [Google Scholar] [CrossRef]
  25. Mahmoodzad, A.; Ghazouani, N.; Mohammed, A.H.; Ibrahim, H.H.; Alghamdi, A.; Albaijan, I.; El Ouni, M.H. Predicting rockbursts in deep tunnels based on ejection velocity and machine learning. Tunn. Undergr. Space Technol. 2024, 166, 105671. [Google Scholar] [CrossRef]
  26. Rabi, R.R.; Monti, G. Genetic algorithm-based model updating in a real-time digital twin for steel bridge monitoring. Appl. Sci. 2025, 15, 4074. [Google Scholar] [CrossRef]
  27. Qian, L.; Pan, Q.; Lv, Y.; Zhao, X. Fault detection of bearing by ResNet classifier with model-based data augmentation. Machines 2022, 10, 521. [Google Scholar] [CrossRef]
  28. Hajihassani, M.; Armaghani, D.J.; Marto, A.; Mohamad, E.T. Ground vibration prediction in quarry blasting through an ANN model optimized with a hybrid of cuckoo search and fireworks algorithms. Eng. Comput. 2021, 37, 331–344. [Google Scholar]
  29. Inderyas, O.; Alver, N.; Tayfur, S.; Shimamoto, Y.; Suzuki, T. Deep learning-based acoustic emission signal filtration model in reinforced concrete. Arab. J. Sci. Eng. 2025, 50, 1885–1903. [Google Scholar] [CrossRef]
  30. Kim, T.H.; Park, J.S. Shear capacity assessment of hollow-core RC piers via machine learning. Eng. Struct. 2025, 300, 117215. [Google Scholar]
  31. Chen, W.; Su, L.; Chen, X.; Huang, Z. Rock image classification using deep residual neural network with transfer learning. Front. Earth Sci. 2023, 10, 1079447. [Google Scholar] [CrossRef]
  32. Xu, Z.; Ma, W.; Lin, P.; Shi, H.; Pan, D.; Liu, T. Deep learning of rock images for intelligent lithology identification. Comput. Geosci. 2021, 154, 104857. [Google Scholar] [CrossRef]
  33. Xing, Y.; Yang, X.; Yu, W. An approach for the classification of rock types using machine learning of core and log data. Sustainability 2023, 15, 8868. [Google Scholar] [CrossRef]
  34. Senjoba, L.; Ikeda, H.; Toriya, H.; Adachi, T.; Kawamura, Y. Deep learning-based rock type identification using drill vibration frequency spectrum images. Int. J. Min. Reclam. Environ. 2025, 39, 40–55. [Google Scholar] [CrossRef]
  35. Zheng, D.; Zhong, H.; Camps-Valls, G.; Cao, Z.; Ma, X.; Mills, B.; Hu, X.; Hou, M.; Ma, C. Explainable deep learning for automatic rock classification. Comput. Geosci. 2024, 184, 105511. [Google Scholar] [CrossRef]
  36. Abdullah, M.A.M.; Mohammed, A.A.; Awad, S.R. RockDNet: Deep learning approach for lithology classification. Appl. Sci. 2024, 14, 5511. [Google Scholar] [CrossRef]
  37. Wang, S.; Ren, Z.; Dong, K.; Li, Y.; Li, J.; Wen, P.; Qu, R.; Li, T.; Wen, Z. Convolutional autoencoder network lithology recognition based on scratch tests. Sci. Rep. 2025, 15, 30401. [Google Scholar] [CrossRef]
  38. Bai, K.; Zhang, Z.; Jin, S.; Dai, S. Rock image classification based on improved EfficientNet. Sci. Rep. 2025, 15, 18683. [Google Scholar] [CrossRef]
  39. Yang, Z.; Cheng, Z.; Wu, D. Deep learning driven prediction and comparative study of surrounding rock deformation in high speed railway tunnels. Sci. Rep. 2025, 15, 24104. [Google Scholar] [CrossRef]
  40. Shu, C.; Meng, Z.; Yang, Y.; Wang, Y.; Liu, S.; Zhang, X.; Zhang, Y. Deep learning-based InSAR time-series deformation prediction in coal mine areas. Geo-Spat. Inf. Sci. 2025, 28, 1–17. [Google Scholar] [CrossRef]
  41. Askaripour, M.; Saeidi, A.; Rouleau, A.; Mercier-Langevin, P. Rockburst in underground excavations: A review of mechanism, classification, and prediction methods. Undergr. Space 2022, 7, 577–607. [Google Scholar] [CrossRef]
  42. Manouchehrian, A.; Cai, M. Prediction of tunnel convergence using machine learning and deep learning approaches: A case study on the Niagara tunnel project. Undergr. Space 2021, 6, 238–251. [Google Scholar]
Figure 1. Structure of depthwise separable convolution.
Figure 1. Structure of depthwise separable convolution.
Sensors 25 07601 g001
Figure 2. RetNet architecture.
Figure 2. RetNet architecture.
Sensors 25 07601 g002
Figure 3. Gated fusion mechanism.
Figure 3. Gated fusion mechanism.
Sensors 25 07601 g003
Figure 4. RDGNet model architecture.
Figure 4. RDGNet model architecture.
Sensors 25 07601 g004
Figure 5. Borehole lithology histogram.
Figure 5. Borehole lithology histogram.
Sensors 25 07601 g005
Figure 6. Brittle failure characteristics of (a) limestone, (b) sandstone, (c) mudstone, and (d) shale.
Figure 6. Brittle failure characteristics of (a) limestone, (b) sandstone, (c) mudstone, and (d) shale.
Sensors 25 07601 g006
Figure 7. Stress and AE energy vs. the time curve.
Figure 7. Stress and AE energy vs. the time curve.
Sensors 25 07601 g007
Figure 8. Mel spectrogram of sample H-01.
Figure 8. Mel spectrogram of sample H-01.
Sensors 25 07601 g008
Figure 9. RDGNet t-SNE feature visualization: (a) features learned through data augmentation; (b) features learned by the sequence branch; (c) features optimized by the gated fusion mechanism.
Figure 9. RDGNet t-SNE feature visualization: (a) features learned through data augmentation; (b) features learned by the sequence branch; (c) features optimized by the gated fusion mechanism.
Sensors 25 07601 g009
Figure 10. RDGNet training accuracy and loss curves.
Figure 10. RDGNet training accuracy and loss curves.
Sensors 25 07601 g010
Figure 11. Validation loss determined from RDGNet ablation experiments.
Figure 11. Validation loss determined from RDGNet ablation experiments.
Sensors 25 07601 g011
Figure 12. RDGNet confusion matrix by rock type.
Figure 12. RDGNet confusion matrix by rock type.
Sensors 25 07601 g012
Figure 13. RDGNet performance comparison by rock type.
Figure 13. RDGNet performance comparison by rock type.
Sensors 25 07601 g013
Figure 14. RDGNet multilithology damage identification using mixed-lithology seismic records.
Figure 14. RDGNet multilithology damage identification using mixed-lithology seismic records.
Sensors 25 07601 g014
Figure 15. RDGNet fusion gate weights (first 30 test samples).
Figure 15. RDGNet fusion gate weights (first 30 test samples).
Sensors 25 07601 g015
Figure 16. Multimodel training accuracy and loss comparison (SNR = 0).
Figure 16. Multimodel training accuracy and loss comparison (SNR = 0).
Sensors 25 07601 g016
Figure 17. Multimodel performance comparison under various noise levels.
Figure 17. Multimodel performance comparison under various noise levels.
Sensors 25 07601 g017
Table 1. Sample distribution across lithologies and data splits.
Table 1. Sample distribution across lithologies and data splits.
SymbolNumber of SpecimensBefore Data
Augmentation
After Data
Augmentation
Training SetValidation SetTest Set
Limestone2027,15350,80927,94412,70310,162
Sandstone15373150,809384117451397
Mudstone10667712,489686931222498
Shale53886496627311242993
Table 2. Summary of data augmentation techniques.
Table 2. Summary of data augmentation techniques.
MethodDescriptionParameters/RangeTarget
Modality
Time stretchingRandomly accelerate or decelerate the sequence λ   ~   U n i f o r m   ( 0.8,1.2 ) AE
sequence
Gaussian noise injectionAdd zero-mean
Gaussian noise
σ _ A E =   0.03 σ _ M e l = 0.003 AE + Mel
spectrogram
Missing-segment simulationRandomly mask contiguous segments of the AE sequence P _ m a s k ~ U n i f o r m   ( 0.10 ,   0.15 ) AE
sequence
Multilithology mixupLinear superposition of 1–4 lithology samples with Dirichlet weights α ~ D i r i c h l e t   ( 1,1 , , 1 ) ( α = 1.0 ) AE + Mel
spectrogram
Table 3. RDGNet ablation results.
Table 3. RDGNet ablation results.
VariantAUCMacro F1AccuracyVal LossPos RecallNeg RecallEpochs
Baseline0.98550.9310.930.0050.930.93230
No Sequence0.93340.89590.8960.00850.890.9025
No Conv0.95280.90620.9050.0060.9050.90728
Enhanced Attention0.95920.9190.9160.00610.890.9530
Replace Focal Loss0.98070.9220.9220.00650.920.92427
No Gated Fusion0.98310.92840.9280.00550.9280.92929
No Augmentation0.89020.76030.760.010.750.7720
Standard RetNet0.96510.90880.920.00580.900.9330
Concatenation Fusion0.95310.90610.900.00570.900.9028
Average Fusion0.95340.89790.910.00540.910.9125
Table 4. Standard RetNet vs. proposed RetNet variant (five independent runs; mean ± standard deviation).
Table 4. Standard RetNet vs. proposed RetNet variant (five independent runs; mean ± standard deviation).
ModelAUCMacro F1AccuracyOptimal
Threshold
Average Time per Round
RDGNet0.9855 ± 0.00410.9310 ± 0.00630.93 ± 0.00690.6914.0 s
RetNet0.9651 ± 0.00360.9088 ± 0.00580.92 ± 0.00550.6916.8 s
Table 5. RDGNet training accuracy for rock type classification.
Table 5. RDGNet training accuracy for rock type classification.
Rock TypeAUCMacro F1RecallPrecision
Limestone0.96420.91240.90870.9178
Sandstone0.95180.88310.87150.8964
Mudstone0.95770.89650.88920.9051
Shale0.97050.92680.92340.9312
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, G.; Ding, H.; Wang, K.; Zhang, X.; Sun, J. Intelligent Recognition of Muffled Blasting Sounds and Lithology Prediction in Coal Mines Based on RDGNet. Sensors 2025, 25, 7601. https://doi.org/10.3390/s25247601

AMA Style

Li G, Ding H, Wang K, Zhang X, Sun J. Intelligent Recognition of Muffled Blasting Sounds and Lithology Prediction in Coal Mines Based on RDGNet. Sensors. 2025; 25(24):7601. https://doi.org/10.3390/s25247601

Chicago/Turabian Style

Li, Gengxin, Hua Ding, Kai Wang, Xiaoqiang Zhang, and Jiacheng Sun. 2025. "Intelligent Recognition of Muffled Blasting Sounds and Lithology Prediction in Coal Mines Based on RDGNet" Sensors 25, no. 24: 7601. https://doi.org/10.3390/s25247601

APA Style

Li, G., Ding, H., Wang, K., Zhang, X., & Sun, J. (2025). Intelligent Recognition of Muffled Blasting Sounds and Lithology Prediction in Coal Mines Based on RDGNet. Sensors, 25(24), 7601. https://doi.org/10.3390/s25247601

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop