You are currently viewing a new version of our website. To view the old version click .
Mathematics
  • Article
  • Open Access

29 May 2024

A Rapid Detection Method for Coal Ash Content in Tailings Suspension Based on Absorption Spectra and Deep Feature Extraction

,
,
,
,
and
1
School of Mechatronic Engineering and Automation, Foshan University, Foshan 528200, China
2
China Coal Technology Engineering Group, Tangshan Research Institute, Tangshan 063000, China
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Mathematics for Visual Computing: Acquisition, Processing, Analysis and Rendering of Visual Information

Abstract

Traditional visual detection methods that employ image data are often unstable due to environmental influences like lighting conditions. However, microfiber spectrometers are capable of capturing the specific wavelength characteristics of tail coal suspensions, effectively circumventing the instability caused by lighting variations. Utilizing spectral analysis techniques for detecting ash content in tail coal appears promising as a more stable method of indirect ash detection. In this context, this paper proposes a rapid detection method for the coal ash content in tailings suspensions based on absorption spectra and deep feature extraction. Initially, a preprocessing method, the inverse time weight function (ITWF), is presented, focusing on the intrinsic connection between the sedimentation phenomena of samples. This enables the model to learn and retain spectral time memory features, thereby enhancing its analytical capabilities. To better capture the spectral characteristics of tail coal suspensions, we designed the DSFN (DeepSpectraFusionNet) model. This model has an MSCR (multi-scale convolutional residual) module, addressing the conventional models’ oversight of the strong correlation between adjacent wavelengths in the spectrum. This facilitates the extraction of relative positional information. Additionally, to uncover potential temporal relationships in sedimentation, we propose a CLSM-CS (convolutional long-short memory with candidate states) module, designed to strengthen the capturing of local information and sequential memory. Ultimately, the method employs a fused convolutional deep classifier to integrate and reconstruct both temporal memory and positional features. This results in a model that effectively correlates the ash content of suspensions with their absorption spectral characteristics. Experimental results confirmed that the proposed model achieved an accuracy of 80.65%, an F1-score of 80.45%, a precision of 83.43%, and a recall of 80.65%. These results outperformed recent coal recognition models and classical temporal models, meeting the high standards required for industrial on-site ash detection tasks.

1. Introduction

Industrial coal slurry flotation is a widely employed physical separation technique for processing various coal waste materials, such as coal gangue and coal slurry. As shown in Figure 1, the primary aim of this method is to enhance the grade of coal (its coal content) and minimize waste by separating coal particles from other impurities in the coal slurry through bubble adsorption. The real-time monitoring of key performance indicators during the production process has increasingly become a focal point of interest in both industrial and academic circles. This heightened attention aims to augment the efficiency of coal slurry flotation technology.
Figure 1. Slime flotation process flow.
Reagents, flotation concentration, and air inflow influence the quality of coal slurry flotation []. Ash content in the tailings is an important indicator for assessing the quality of floated coal and serves as a crucial basis for the rational application of flotation coal. If the ash content in the tailings is low, this indicates that the flotation process has effectively separated impurities and ash components from coal particles, thereby improving the coal’s grade. This usually means that the flotation technology is efficient and the coal quality is better, leading to a higher value in subsequent industrial applications. Conversely, a high tailings ash content indicates poor separation in the flotation process, where impurities and ash components are inadequately separated from coal particles. This may be due to poor bubble adsorption, improper use of chemical reagents, or insufficient bubble generation, among other reasons. A high tailings ash content can reduce the coal’s grade and, consequently, lower its industrial usability.
Traditional methods for ash content detection in coal slurry flotation mainly include radiation measurement and image processing. The radiation measurement method uses the attenuation of X-rays in coal to measure the content of high atomic number elements in coal, from which ash content can be estimated. This non-contact method is suitable for online detection in special environments such as high temperature and pressure. However, it involves using radioactive sources and carries certain safety risks. On the other hand, image processing uses machine vision techniques to extract image features from the surface of coal or foam layers. Then, it establishes a relationship model between image features and ash content to estimate the ash content. This method can capture information about coal particle shapes, colors, and other features to a certain extent through machine learning, providing more potential information. However, it can only measure the surface ash content of flotation and cannot reflect the distribution of internal ash content. Moreover, environmental lighting conditions severely affect the image quality and detection accuracy, requiring appropriate calibration and control.
Therefore, we compiled a comparison of the advantages and disadvantages of the main research objects currently in use, as shown in Table 1. Currently, the detection of ash content in coal slurry flotation tailings typically involves collecting images of dried coal powder or flotation froth as datasets. However, these methods all share the limitation of being heavily influenced by the quality of the images, which is significantly affected by environmental lighting conditions. It is difficult to rigorously control lighting factors in industrial environments. As for the reflectance spectra, they are sampled based on the reflective properties of substances to light. However, fundamentally, they still use dried coal powder as the research object, which has the following drawbacks: Firstly, during the coal slurry flotation process, tail coal is usually in a suspended state. To obtain dried coal powder, drying is required, which takes tens of minutes to several hours to complete, not meeting industrial requirements and entailing significant delays. Secondly, the above methods can only obtain surface information of samples and cannot reflect deep internal information. To address these issues, we considered using absorption spectra of tailings suspension as the research object. The benefits of this approach are as follows: Firstly, it directly eliminates the drying step, greatly shortening the detection time and meeting rapid industrial requirements. Secondly, absorption spectra can reflect the dynamic changes in suspended liquid samples, containing rich feature information. To address the difficulty of capturing these features, we considered using deep learning methods to fully explore deep information, relying on its powerful modeling capabilities. In summary, this paper proposes a rapid detection method for coal ash content in tailings suspensions based on absorption spectra and deep feature extraction. The spectral data of tail coal suspensions were preprocessed using the inverse time weight preprocessing method. The DSFN model was designed to extract location and time information from the spectral data using an MSCR module and CLSM-CS module. Finally, the features extracted by the two modules were fused and reconstructed to output information, thereby constructing a correlation model between the ash content of the suspension and the absorption spectral characteristics. The contributions of this paper are as follows:
Table 1. Comparison of different research objects for coal slime flotation ash content detection methods.
  • Our method focuses on the absorption spectrum of tailings suspensions, marking the first study in the field to detect ash content in coal slurry flotation tailings. This approach not only meets the demands for industrial rapidity but also provides more comprehensive data.
  • Traditional sequential preprocessing methods fail to effectively capture the temporal relationships inherent in the settling phenomena of tailings suspension samples. To address this, we propose an inverse time weight function (ITWF) that emphasizes differences at earlier time points, while still considering information from later time points.
  • Considering the unique characteristics of the absorption spectrum of tailings suspension, we designed the DSFN (DeepSpectraFusionNet) model. This model effectively captures the location-dependent features and time memory features of the tailings suspension absorption spectrum, yielding promising results.

3. Materials and Methods

3.1. Data Description and Preprocessing

The experimental equipment required for building the data acquisition system using the absorption spectral data of tail coal suspensions included a quartz cuvette, a magnetic stirrer, a fiber-optic spectrometer, a UV quartz fiber optic, a dual-path cuvette holder, a pulsed xenon lamp light source, and so on. From Figure 2 and Figure 3, we can understand the data acquisition process. First, the xenon lamp light source instrument outputs light, which is then received by a collimating mirror and passes through a cuvette sample. Afterward, the light is output by another collimating mirror and received by a miniature fiber-optic spectrometer. The absorbance curve is displayed on the screen.
Figure 2. Experimental data collection site.
Figure 3. Spectrum acquisition flow chart.
The experimental samples consisted of tail coal samples with ash contents of 25.6%, 34.4%, 38.0%, 44.0%, and 49.4%. The corresponding ash content labels for these samples were accurately determined using the fast ashing experiment commonly employed in the industry. Each tail coal sample was mixed with an equal amount of pure water to create a suspension, simulating the state of tail coal extraction in an industrial setting. Samples were obtained by taking a portion of the suspension from each sample. Since the samples were in suspension, settling phenomena occurred over time. To mitigate the impact on the experimental results, we ran a magnetic stirrer for 5 min at the beginning of each experiment to ensure thorough mixing. The cuvette holder used had a hollow structure, allowing us to collect absorbance spectra immediately at 0 s after stopping the magnetic stirring. Then, the samples were allowed to settle, and the absorbance spectra were collected at 0 s, 30 s, 60 s, 90 s, and 300 s. For each category, 31 samples were taken at each time point, resulting in five ash content categories. Therefore, a total of 775 data samples were obtained. The data type is a one-dimensional univariate sequence containing 2088 feature points. Figure 4 shows spectral data curves for 30 randomly selected samples from each category. Each curve represents a spectral data sample, illustrating the general trend of the spectral data for each category.
Figure 4. Thirty randomly selected spectral data samples display.

3.1.1. Data Preprocessing

Preprocessing is widely executed when analyzing spectral signals, to eliminate or reduce unnecessary variations in the original spectral data, such as noise bands and outliers. Moreover, appropriate preprocessing is crucial for enhancing the accuracy of neural networks [].
Preprocessing included three steps. Firstly, second-order polynomial Savitzky–Golay (SG) smoothing denoised the input data’s raw absorbance spectra. Savitzky–Golay smoothing is a technique used to smooth data by fitting local data with a polynomial, which helps reduce noise and outliers, while preserving the data’s main features. Its formula is as follows:
Y i = j = k k c j · Y i + j
where Y i represents the smoothed data, i indicates the index of the data points in the series, k is the order of the polynomial fitted, and c j are the Savitzky–Golay coefficients.
Secondly, first-order derivative correction was utilized to correct baseline drift. This step aimed to correct the baseline drift, which is the gradual increasing or decreasing trend that might exist in a spectrum. By calculating the data’s first-order derivative (difference), baseline drift can be eliminated, thereby better highlighting the absorption peaks in the spectrum. Its formula is as follows:
D i = Y i + 1 Y i
Y i = Y i mean ( D )
where D i is the first-order derivative of the original data, Y i is the corrected data point after baseline drift correction, and m e a n ( D ) represents the mean value of the first-order derivative.
Z i = Y i min Y i max Y i min Y i
where max Y i and min Y i are the minimum and maximum values of the corrected data, respectively.
This benefits many machine learning algorithms by reducing the amplitude differences between different features, aiding the model in better convergence and prediction. The processed data are shown in Figure 4.

3.1.2. Inverse Time Weight Function

Beyond the usual spectrum preprocessing methods, the data analysis also had to account for the settling phenomenon of the suspended liquid samples. As illustrated in Figure 5, particles within the suspension generally possessed higher densities than the liquid medium, causing them to gradually settle or precipitate over time. Figure 6 showcases the evolution of the absorption spectra over different time intervals, displaying a discernible, orderly gradient-like transition from 0 to 300 s.
Figure 5. Sedimentation phenomenon of tail coal.
Figure 6. Changes in spectral data at different times.
Based on this, we proposed an inverse time weight function (ITWF). At moment 0, the sample data contain the richest information regarding differences. As time progresses, the coal tailings particles continually settle to the bottom and come to rest, making the absorption spectrum at this stage similar to that of pure water, containing the least amount of information. A weight decay factor emphasizes data from earlier time points, focusing on the differences at these points, while not ignoring information from later times. Data collection was performed at different time points for the same sample, and a temporal memory rule was used between the corresponding absorption spectra. These samples were further processed into temporal memory data, enabling the subsequent models to better mine temporal feature information. The specific formula is as follows:
Y t = i = 1 n α i x i t i = 1 n α i
where x i represents the absorption spectrum sequence at different time points, and α i is the weight decay factor.

3.2. Model Framework

Building upon in-depth research into coal slime flotation ash content detection, this paper proposed a rapid detection method for coal ash content in tailings suspension based on absorption spectra and deep feature extraction. The overall framework of the DSFN is shown in Figure 7.
Figure 7. Overall framework structure.
Following the data collection and preprocessing methods described above, the data were processed into an original dataset and a temporary dataset. To address the distinct characteristics of these datasets, we designed an MSCR module and CLSM-CS module to extract specific features. After completing the information extraction for each module, a fused convolutional deep classifier was employed to merge and reconstruct the dual information. This approach enabled the construction of a model that correlated the ash content in the suspension with its absorption spectral characteristics for ash detection.

3.2.1. Multi-Scale Convolutional Residual Module

As shown in Figure 8 and Figure 9, we have used two types of figures to make the description of this module clearer. In the diagram, [1 × 3]:14 indicates a convolutional kernel size of 1 × 3 with 14 output channels, and so on. To address the differentiated features of spectral data at various wavelengths, this paper designed an MSCR module based on conventional convolutional neural networks. Traditional convolutional neural networks typically use fixed-size convolution kernels, which may result in a limited receptive field and difficulty in effectively capturing features of different scales and levels in the input data. Additionally, conventional convolution operations only extract information within a local receptive field, potentially making it difficult for the model to understand long-distance dependencies between pixels. By introducing convolution kernels of multiple scales, the model can better process features of different scales. Smaller kernels can capture finer details, while larger ones can capture broader contextual information. Conventional multi-scale convolutions usually just concatenate information of different scales, potentially overlooking the contextual relationships between scales and generating redundant information. To address these issues, residual connections and two-dimensional convolutional concatenation were introduced. Residual connections provide a direct shortcut bypassing intermediate layers, establishing long-distance information dependencies, allowing for faster information propagation and enabling the model to learn residual features and the differences between the input and output. Concatenating features of different scales and using two-dimensional convolution for feature extraction avoids redundancy, while reducing computational and memory overheads. The specific steps are as follows:
Figure 8. MSCR Module Visualization Flowchart.
Figure 9. MSCR Module Structure Diagram.
First, the data are processed with embedding, to convert the discrete wavelength data into continuous, dense vector representations. The formula is as follows:
P E = p e p o s , 2 i = sin p o s 10 , 000 2 i d model p e pos , 2 i + 1 = cos p o s 10 , 000 2 i d model
V E = Conv 1 d X b
X e m b = sum ( V E + P E )
where P E represents position embedding, p o s is the position in the input sequence, d m o d e l is the dimension of the model, i ( 0 , d m o d e l 2 1 ) , and V E represents numerical embedding.
Then, a set of convolution kernels of different scales was designed to extract features from the data. This enables the model to more comprehensively understand time-series data and learn features and patterns at different window scales. The specific steps are as follows:
X i = Norm Dropout Tanh Conv1d x emb kernel = i
where i is the index of the convolution kernel, corresponding to different kernel sizes for convolution.
Then, upsampling and truncation operations are performed to match the sequence length for subsequent processing. The features extracted are added to the original output to form a residual connection, with the specific steps as follows:
Y i = Truncate Norm Dropout Tanh ConvTranspose1d X i kernel = i
Z i = X e m b + Y i
After processing through the MSCR module, the feature representations processed under different sliding windows are stored in a list for subsequent processing. This list contains features at multiple scales, which are expanded by one dimension. The features of several scales are concatenated in the newly added dimension according to the number of multi-scale convolution kernels. The resulting new features, rearranged in dimensions to resemble a two-dimensional tensor like an image, are processed using common two-dimensional convolution operations for multi-scale concatenation. The steps are as follows:
Z total = torch · cat i = 0 len ( kernels ) 1 Z i · unsqueeze ( 2 ) , dim = 2
M = Linear Conv 2 d Z total · permute ( 0 , 3 , 1 , 2 )
where i is the index of the convolution kernel. The final output M is used as F e a t u r e M and is subsequently fused with the final output F e a t u r e C from the CLSM-CS module.

3.2.2. Convolutional Long-Short Memory with Candidate States Module

Considering the sedimentation phenomenon in suspension samples, as shown in Figure 5 and Figure 6, the absorbance spectra of the same sample differ at various time points, meaning that the spectral information of suspension samples has temporal memory continuity during the sedimentation process. Although the classic LSTM (long short-term memory network) is undoubtedly one of the best choices, its direct application to the extraction of spectral features still faces two issues: Firstly, LSTM mainly models the global information of the entire sequence. At the same time, the temporal memory dataset tends to capture local time information. Secondly, while LSTM effectively captures long-term dependencies, it may overly focus on past information, neglecting the input at the current moment, which is one of the most important inputs for the suspension absorbance spectrum []. The CLSM-CS module was designed to address the limitations of traditional LSTM in processing sediment spectral data. By incorporating convolutional operations, CLSM-CS can more effectively capture local features in time series data and enhance the model’s sensitivity to current inputs. In analyzing sedimentation phenomena, spectral information at various time points indicates the dynamic changes in the sample, and the unique structure of CLSM-CS optimizes the processing of this information. Specifically, changes in spectral information at different time points are directly reflected in the variations in feature point values corresponding to the same wavelengths. Each time step in CLSM-CS involves computations through a series of convolutional layers. The convolution operations sensitively capture these changes in feature points, while also updating the cell state and directly influencing the activation of gate controllers. This allows the model to adaptively adjust its internal state at each time point, to reflect the latest input changes. This mechanism gives CLSM-CS a significant advantage in analyzing temporal correlations in the sedimentation process, particularly in identifying critical moments of change.
As shown in Figure 10, for each time step and CLSM-CS layer, the specific formulas of the CLSM-CS method are as follows:
Figure 10. CLSM-CS module.
Firstly, utilize convolutional operations to simulate the computation of four gate structures.
G t ( l ) = Conv1d X t ( l 1 )
where G t ( l ) represents the convolutional output, and X t ( l 1 ) is the input from the previous layer or the original input (when l = 1 ). This step uses a convolutional kernel size of 3 and an output channel count of 16.
Subsequently, expand into four parts to simulate four gate structures:
G t ( l ) = I t ( l ) , F t ( l ) , C ˜ t ( l ) , O t ( l )
Next, apply activation functions to simulate the updating of gates:
i t ( l ) = σ I t ( l ) f t ( l ) = σ F t ( l ) c ˜ t ( l ) = tanh C ˜ t ( l ) o t ( l ) = tanh o t ( l )
Then, update the cell state and hidden state:
c t ( l ) = f t ( l ) c t 1 ( l ) + i t ( l ) c ˜ t ( l )
h t ( l ) = o t ( l ) tanh c t ( l )
where ⊙ represents the Hadamard product, σ is the sigmoid activation function, tanh represents the hyperbolic tangent activation function, c t ( l ) is the cell state at time step t of layer l, and h t ( l ) is the output of layer l at time step t, also used as the input for the next layer.
Lastly, for the final output layer:
C = Conv1d h t ( last )
Here, the convolutional kernel size is 1, and the output channel count is 7. C is the output of the last CLSM-CS layer and is the final output of the network. It is used as F e a t u r e C and subsequently fused with F e a t u r e M , which is extracted by the MSCR module. In this method, a single layer of CLSM-CS is used.

3.2.3. Fusion Convolutional Deep Classifier

After separately extracting F e a t u r e M and F e a t u r e C , this paper used a fusion convolutional deep classifier to address how to effectively integrate both sets of features and mine their interrelated information. As shown in Figure 11 and Figure 12, we use two types of figures to make the description of this module clearer. In the diagram, [3 × 3]:7 indicates a convolutional kernel size of 3 × 3 with 7 output channels. To introduce feature fusion, we expand the outputs of both convolutions by appending a vector of one at the end of the features.
M = unsqueeze Position output , 2
C = unsqueeze Time output , 1
where M represents the expanded position-dependent feature, and C is the expanded temporal memory feature.
Figure 11. Fusion convolutional deep classifier visualization flowchart.
Figure 12. Fusion convolutional deep classifier structure diagram.
Then, by specifying the index relationship between input and output, the two corresponding feature tensors are multiplied and summed, thus achieving feature fusion.
Fusion output = einsum M , C
Further feature compression and extraction are carried out through a two-dimensional convolution layer, and the features are then flattened. The convolutional kernel size is 3, and the output channel count is 7.
Output = flatten Conv 2 d Fusion output , dim = 1
Finally, through a fully connected layer and a projection layer, we map the abstracted features to the final classification results.
Final output = projection ( linear ( Output ) )
Both sets of feature information can be fully utilized through the above method, without loss. Furthermore, the fusion of feature tensors uncovers potential connections between the two types of features, effectively capturing the input data’s complex structure and associated information.

4. Experiment

All experiments were conducted in the PyTorch 1.13.1 + CUDA 11.6 framework using an NVIDIA RTX 4090 GPU, with code implemented in Python 3.9.16. Table 2 shows some of the parameters used for model training.
Table 2. Parameters Used for Model Training.

4.1. Comparison of Preprocessing Methods

To verify the effectiveness of the ITWF, other common sequence processing methods were used for comparison. These methods included global averaging, and sliding window averaging. The original data without any function processing and data processed using a downstream time-weighted function were also compared to further validate the effectiveness. The specific formulas for each method are as follows:
Global average (GA) method, which involves directly averaging data across different time points.
Y t = 1 m i = 1 m x i
where m is the number of absorbance spectra, and x i represents the absorbance spectrum sequence at different time points.
Sliding window average (SWA) method, which uses a specified window size to slide through the values for averaging.
Y t = 1 m j = 1 m 1 w k = i w + 1 i x i
where w is the size of the sliding window, j is the index of the corresponding sequence, and k is the index of elements in the sequence.
Downstream time-weighted function (DTF) method, which differs from the method in this paper by focusing on the spectral information at later time points.
Y t = i = 1 n α n i x i i = 1 n α n i
Original dataset (OD) method, without any defined function processing.
Table 3 and Figure 13 present the detection results of the methods. The experimental results showed that our method (ITWF) performed the best, followed by Method B (sliding window average), then Method A (global average), followed by Method D (original dataset without defined function processing), and finally Method C (downstream time-weighted function). Our method was the most effective in dealing with the sedimentation phenomenon in suspension samples. Because the ITWF emphasized data from earlier time points, it better captured the differential information of the suspension samples at the initial moment. This method excels in handling the richness of information in dynamic processes. The poorer performance of the other methods can be attributed to their varying degrees of neglect for the dynamic nature of spectral feature changes.
Table 3. Comparison of five preprocessing methods.
Figure 13. Preprocessing method visualization.

4.2. Model Comparison Experiment

Additionally, we evaluated several typical methods of coal identification [], which included random forest (RF), support vector machine (SVM), extreme learning machine (ELM), tensor extreme learning machine (TELM), PCA-ELM (principal component analysis for extreme learning machine), and convolutional neural network with extreme learning machine (CNN + ELM) methods. The CNN + ELM method was adapted from []. Originally designed for two-dimensional image data, we modified it to accommodate one-dimensional data dimensions. Figure 14 illustrates its network structure. In the diagram, [1 × 3]:32 indicates that the convolutional kernel size was 1 × 3 with 32 output channels, and so on. Furthermore, we compared some representative time-series networks: LSTM [], TCN (temporal convolutional network) [], and Transformer []. We evaluated each method based on four key performance metrics: accuracy, F1 score, precision, and recall, to validate the superiority of the model proposed in this paper.
Figure 14. CNN + ELM network architecture.
As shown in Table 4 and Figure 15, firstly, from an overall perspective, significant differences were observed in the performance of the various methods across the four evaluation metrics: accuracy, F1 score, precision, and recall. These metrics are key factors in measuring the performance of machine learning models and collectively determine a model’s overall effectiveness. Our approach demonstrated the best performance across all four metrics among these methods. RF and SVM performed relatively poorly, indicating their limited adaptability to the data. TELM and ELM exhibited a similar performance, with TELM showing slight improvements in precision and recall. PCA_ELM underperformed compared to ELM, suggesting that PCA did not significantly enhance ELM’s performance in this context. The LSTM model’s performance was average, particularly lacking in accuracy and F1-score. The TCN model showed good performance in terms of precision. Both the CNN + ELM and Transformer models performed well across all metrics.
Table 4. Comparison of model results.
Figure 15. Performance metric visualization.

4.3. Ablation Study

To verify the effectiveness of the improved MSCR module and CLSM-CS module, we conducted ablation experiments, as shown in Table 5. In these experiments, DSFN-A represents the model without the MSCR module (replaced by a single convolutional layer with a kernel size of 3), DSFN-B represents the model without the CLSM-CS module (replaced by LSTM with the same input parameters), and DSFN-AB indicates the model with both modules removed. It is evident that the MSCR module and the CLSM-CS module were effective in improving the model performance, with the best results achieved through their synergistic interaction.
Table 5. Ablation Results.

4.4. Performance Evaluation of Each Category

Lastly, we also examined the model’s performance in individual categories rather than overall effectiveness. Figure 16 presents a confusion matrix, commonly used for evaluating classification models. The horizontal axis denotes the model’s predicted labels, while the vertical axis represents the true labels. The intensity of colors reflects sample quantities, with darker shades indicating higher counts. It is evident that the model accurately classified the majority of samples. In Figure 17, we depict the recall and precision for each category separately, highlighting the model’s most precise detection for the 34.4% category. However, the model exhibited a tendency to confuse similar categories, as illustrated in Figure 15, where a significant proportion of samples, originally labeled as 49.4%, were misclassified as 44.0%. Addressing this limitation represents one of the areas for future improvement.
Figure 16. Confusion matrix of model prediction.
Figure 17. Prediction of different ash contents.

5. Discussion

Our method presents a tailored approach by directing its focus towards analyzing the absorption spectra of coal slurry flotation tailings. To tackle the sedimentation issue observed in samples, it puts forward the concept of ITWF. This innovative function aims to highlight the characteristics of early-time data, while retaining valuable late-time information. Additionally, we designed the DSFN (DeepSpectraFusionNet) model to specifically extract features from our data. The experimental findings further underscored the method’s efficacy and superiority. This marks the first proposal in the field of detecting ash content in coal slurry flotation tailings, offering a fresh perspective for research.
Limitations: Although this study demonstrated certain advantages compared to conventional methods, from data selection to model design, we must also acknowledge its limitations. Firstly, concerning data selection, the presence of sedimentation phenomena necessitated a sampling duration of at least approximately 5 min per data point, which still falls short of the ideal requirements for rapid industrial testing. Then, while particles in the suspended liquid samples tended to settle over time, they still exhibited random movements, and capturing information about the random motion of sample particles while discarding redundant information is a major challenge in model design. Additionally, the industrial setting requires smaller and faster model structures, and our model still has some gaps in meeting these needs. To address these fundamental issues, future research will focus on redesigning and optimizing the sampling platform, to achieve faster, smaller, and more accurate detection methods.

6. Conclusions

Addressing the stability issues associated with traditional image detection methods, this paper focused on the state of flotation tailings coal suspensions as the research subject. Utilizing a microfiber-optic spectrometer to collect absorbance spectrum data reduced the dependency on external conditions and enhanced industrial adaptability. A rapid detection method for coal ash content in tailings suspension based on absorption spectra and deep feature extraction was proposed. The ITWF emphasized the inherent connection of sample sedimentation, enhancing the model’s ability to learn spectral temporal memory features. We designed a DSFN model to specifically extract relevant features. Using the MSCR module enabled the model to extract strongly correlated positional information between adjacent wavelengths in the spectrum. In contrast, introducing the CLSM-CS module strengthened the capturing of local information and temporal memory. Finally, applying a fusion convolutional deep classifier effectively merged temporal memory features with position-dependent features, creating a precise ash content detection correlation model. The experimental results demonstrated the method’s high stability, accuracy, and rapidity in industrial field applications, showcasing its immense potential in tailings coal ash content detection.

Author Contributions

Conceptualization: W.Z. and X.Z.; methodology: W.Z. and X.Z.; software: W.Z. and X.Z.; validation: X.Z., N.L. and Z.Z. (Zhengquan Zhang); formal analysis: W.F., N.L. and Z.Z. (Zhengquan Zhang); investigation: Z.Z. (Zhengjun Zhu); resources: Z.Z. (Zhengjun Zhu); data curation: W.F. and N.L.; writing—original draft preparation, W.Z. and X.Z.; writing—review and editing, W.Z. and X.Z.; project administration: W.Z. and Z.Z. (Zhengjun Zhu); funding acquisition: W.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China grant number 62106048; in part by the Key Research and Development Program of Guangdong, China grant numbers 2020B0404030001 and 2021B010410002.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Liang, H.; Li, Y. Analyses of influencing factors of fine coal flotation. Coal Process. Compr. Util. 2023, 20–24. [Google Scholar] [CrossRef]
  2. Fookes, R.; Gravitis, V.; Watt, J.; Hartley, P.; Campbell, C.; Howells, E.; McLennan, T.; Millen, M. On-line determination of the ash content of coal using a “Siroash” gauge based on the transmission of low and high energy γ-rays. Int. J. Appl. Radiat. Isot. 1983, 34, 63–69. [Google Scholar] [CrossRef]
  3. Vardhan, R.H.; Giribabu. On-Line Coal-Ash Monitoring Technologies in Coal Washaries—A Review. Procedia Earth Planet. Sci. 2015, 11, 49–55. [Google Scholar] [CrossRef]
  4. Li, J.; Zhang, J.; Ge, L.; Zhou, W.; Zhong, D. Software design method and application of Monte Carlo simulation of NaI detector natural γ spectrum. Nucl. Electron. Detect. Technol. 2005, 423–425. [Google Scholar]
  5. Huang, X.; Wang, G.; Sun, P.; Yang, D.; Ma, Y. Low-energy γ-ray backscatter method for measuring coal ash. Nucl. Tech. 2005, 72–75. [Google Scholar]
  6. Bona, M.; Andrés, J. Coal analysis by diffuse reflectance near-infrared spectroscopy: Hierarchical cluster and linear discriminant analysis. Talanta 2007, 72, 1423–1431. [Google Scholar] [CrossRef]
  7. Gui, W.; Yang, C.; Xu, D.; Lu, M.; Xie, Y. Research progress on mineral flotation process monitoring technology based on machine vision. Acta Autom. Sin. 2013, 39, 1879–1888. [Google Scholar] [CrossRef]
  8. Zhao, B.; Hu, S.; Zhao, X.; Zhou, B.; Li, J.; Huang, W.; Chen, G.; Wu, C.; Liu, K. The application of machine learning models based on particles characteristics during coal slime flotation. Adv. Powder Technol. 2022, 33, 103363. [Google Scholar] [CrossRef]
  9. Tan, L.; Zhang, Z.; Zhao, W.; Zhang, R.; Zhang, K. Research progress on mineral flotation foam monitoring based on machine vision. Min. Res. Dev. 2020, 40, 123–130. [Google Scholar] [CrossRef]
  10. Wang, J.; Wang, R.; Fu, X.; Wu, T. Research on soft measurement of flotation tail coal ash based on color image processing. Coal Eng. 2020, 52, 137–142. [Google Scholar]
  11. Ding, J.; Bai, F.; Ren, X.; Fan, M. Ash content prediction system of clean coal based on flotation foam image recognition. Coal Prep. Technol. 2022, 50, 89–93. [Google Scholar] [CrossRef]
  12. Guo, X.; Wei, L.; Yang, C. Study on coal slime flotation tailings ash detection method based on deep convolutional network. Coal Technol. 2020, 39, 144–146. [Google Scholar] [CrossRef]
  13. Wen, Z.; Zhou, C.; Pan, J.; Nie, T.; Zhou, C.; Lu, Z. Deep learning-based ash content prediction of coal flotation concentrate using convolutional neural network. Miner. Eng. 2021, 174, 107251. [Google Scholar] [CrossRef]
  14. Yang, X.; Zhang, K.; Ni, C.; Cao, H.; Thé, J.; Xie, G.; Tan, Z.; Yu, H. Ash determination of coal flotation concentrate by analyzing froth image using a novel hybrid model based on deep learning algorithms and attention mechanism. Energy 2022, 260, 125027. [Google Scholar] [CrossRef]
  15. Han, Y.; Wang, L.; Liu, Q.; Gui, X. Intelligent detection method of flotation tail coal ash based on CNN-BP. Ind. Min. Autom. 2023, 49, 100–106. [Google Scholar] [CrossRef]
  16. Manley, M. Near-infrared spectroscopy and hyperspectral imaging: Non-destructive analysis of biological materials. Chem. Soc. Rev. 2014, 43, 8200–8214. [Google Scholar] [CrossRef] [PubMed]
  17. Yang, J.; Wang, J.; Lu, G.; Fei, S.; Yan, T.; Zhang, C.; Lu, X.; Yu, Z.; Li, W.; Tang, X. TeaNet: Deep learning on Near-Infrared Spectroscopy (NIR) data for the assurance of tea quality. Comput. Electron. Agric. 2021, 190, 106431. [Google Scholar] [CrossRef]
  18. Chang, Y.T.; Hsueh, M.C.; Hung, S.P.; Lu, J.M.; Peng, J.H.; Chen, S.F. Prediction of specialty coffee flavors based on near-infrared spectra using machine-and deep-learning methods. J. Sci. Food Agric. 2021, 101, 4705–4714. [Google Scholar] [CrossRef] [PubMed]
  19. Li, X.; Fan, P.; Li, Z.; Chen, G.; Qiu, H.; Hou, G. Soil classification based on deep learning algorithm and visible near-infrared spectroscopy. J. Spectrosc. 2021, 2021, 1508267. [Google Scholar] [CrossRef]
  20. Xiao, D.; Vu, Q.H.; Le, B.T. Salt content in saline-alkali soil detection using visible-near infrared spectroscopy and a 2D deep learning. Microchem. J. 2021, 165, 106182. [Google Scholar] [CrossRef]
  21. Liu, J.; Zhang, J.; Tan, Z.; Hou, Q.; Liu, R. Detecting the content of the bright blue pigment in cream based on deep learning and near-infrared spectroscopy. Spectrochim. Acta Part A Mol. Biomol. Spectrosc. 2022, 270, 120757. [Google Scholar] [CrossRef] [PubMed]
  22. Bai, X.; Zhang, L.; Kang, C.; Quan, B.; Zheng, Y.; Zhang, X.; Song, J.; Xia, T.; Wang, M. Near-infrared spectroscopy and machine learning-based technique to predict quality-related parameters in instant tea. Sci. Rep. 2022, 12, 3833. [Google Scholar] [CrossRef] [PubMed]
  23. Ravichandran, P.; Viswanathan, S.; Ravichandran, S.; Pan, Y.J.; Chang, Y.K. Estimation of grain quality parameters in rice for high-throughput screening with near-infrared spectroscopy and deep learning. Cereal Chem. 2022, 99, 907–919. [Google Scholar] [CrossRef]
  24. Lee, P.H.; Torng, C.C.; Lin, C.H.; Chou, C.Y. Control chart pattern recognition using spectral clustering technique and support vector machine under gamma distribution. Comput. Ind. Eng. 2022, 171, 108437. [Google Scholar] [CrossRef]
  25. Ma, Y.; Lan, Y.; Xie, Y.; Yu, L.; Chen, C.; Wu, Y.; Dai, X. A Spatial–Spectral Transformer for Hyperspectral Image Classification Based on Global Dependencies of Multi-Scale Features. Remote Sens. 2024, 16, 404. [Google Scholar] [CrossRef]
  26. Le, B.T.; Xiao, D.; Mao, Y.; He, D. Coal analysis based on visible-infrared spectroscopy and a deep neural network. Infrared Phys. Technol. 2018, 93, 34–40. [Google Scholar] [CrossRef]
  27. Mao, Y.; Le, B.T.; Xiao, D.; He, D.; Liu, C.; Jiang, L.; Yu, Z.; Yang, F.; Liu, X. Coal classification method based on visible-infrared spectroscopy and an improved multilayer extreme learning machine. Opt. Laser Technol. 2019, 114, 10–15. [Google Scholar] [CrossRef]
  28. Xiao, D.; Le, B.T. Rapid analysis of coal characteristics based on deep learning and visible-infrared spectroscopy. Microchem. J. 2020, 157, 104880. [Google Scholar] [CrossRef]
  29. Xiao, D.; Le, T.T.G.; Doan, T.T.; Le, B.T. Coal identification based on a deep network and reflectance spectroscopy. Spectrochim. Acta Part A Mol. Biomol. Spectrosc. 2022, 270, 120859. [Google Scholar] [CrossRef]
  30. Xiao, D.; Yan, Z.; Li, J.; Fu, Y.; Li, Z. Rapid proximate analysis of coal based on reflectance spectroscopy and deep learning. Spectrochim. Acta Part A Mol. Biomol. Spectrosc. 2023, 287, 122042. [Google Scholar] [CrossRef]
  31. Li, B.; Xiao, D.; Xie, H.; Huang, J.; Yan, Z. Coal Classification Based on Reflection Spectroscopy and the IAT-TELM Algorithm. ACS Omega 2023, 8, 35232–35241. [Google Scholar] [CrossRef]
  32. Li, Z.; Wang, D.; Zhu, T.; Ni, C.; Zhou, C. SCNet: A deep learning network framework for analyzing near-infrared spectroscopy using short-cut. Infrared Phys. Technol. 2023, 132, 104731. [Google Scholar] [CrossRef]
  33. Hua, Y.; Zhao, Z.; Li, R.; Chen, X.; Liu, Z.; Zhang, H. Deep learning with long short-term memory for time series prediction. IEEE Commun. Mag. 2019, 57, 114–119. [Google Scholar] [CrossRef]
  34. Yu, Y.; Si, X.; Hu, C.; Zhang, J. A review of recurrent neural networks: LSTM cells and network architectures. Neural Comput. 2019, 31, 1235–1270. [Google Scholar] [CrossRef] [PubMed]
  35. Bai, S.; Kolter, J.Z.; Koltun, V. An empirical evaluation of generic convolutional and recurrent networks for sequence modeling. arXiv 2018, arXiv:1803.01271. [Google Scholar]
  36. Lin, T.; Wang, Y.; Liu, X.; Qiu, X. A survey of transformers. AI Open 2022, 3, 111–132. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.