You are currently viewing a new version of our website. To view the old version click .
Applied Sciences
  • Article
  • Open Access

31 March 2024

Research on Large Language Model for Coal Mine Equipment Maintenance Based on Multi-Source Text

,
,
,
and
1
School of Mechanical Engineering, Xi’an University of Science and Technology, Xi’an 710054, China
2
Shaanxi Key Laboratory of Intelligent Detection and Control of Mechanical and Electrical Equipment, Xi’an 710054, China
*
Author to whom correspondence should be addressed.

Abstract

The efficient management and utilization of coal mine equipment maintenance knowledge is an indispensable foundation for advancing the establishment of intelligent mines. This knowledge has problems such as scattered, low sharing, and insufficient management, which restricts the development of coal mine intelligence. For the above-mentioned problems, a large language model for the maintenance of coal mine equipment based on multi-source text (XCoalChat) was proposed to better manage and utilize the existing massive knowledge of coal mine equipment maintenance. The dataset of coal mine equipment maintenance based on ReliableCEMK-Self-Instruction was constructed to obtain a wide and diverse amount of knowledge through sample generation. Aiming at the illusory problem of the large language model, a knowledge graph enhancement method based on the “Coal Mine Equipment Maintenance System—Full Life Cycle—Specification” was proposed to improve the knowledge density. A triple-LoRA fine-tuning mechanism and DPO direct preference optimization method were introduced into the top of the baseline model, which guarantees that XCoalChat can handle multiple Q&A and maintenance decision analysis tasks with limited computing power. Compared with ChatGLM, Bloom, and LLama, the comprehensive assessment of XCoalChat was performed by experiments including coal mine dialog consulting, coal mine professional consulting, and maintenance decision analysis. The results showed that XCoalChat achieved the best response accuracy in professional consulting and maintenance decision analysis; XCoalChat also took the least reasoning time on average. XCoalChat outperformed other mainstream large language models, which verify that XCoalChat is an effective large language model in the field of coal mine equipment maintenance.

1. Introduction

The “Guiding Opinions on Accelerating the Intelligent Development of Coal Mines” point out that it is necessary to accelerate the supply-side structural reform and high-quality development of the coal industry, which is a milestone for the development of China’s coal industry [1]. With the ongoing development of intelligent technology in coal mines, the degree of informatization and intelligence of coal mine equipment has also been constantly improved. Intelligent mines are trying to introduce big data analysis and other technologies, with the key equipment of the comprehensive release working face as the research object. It forms the multi-source heterogeneous big data-driven intelligent identification of the fault state of the comprehensive coal mine equipment and the fault prediction technology system through multi-source heterogeneous data collection and the cleaning and distribution storage [2], the state identification of the equipment of the comprehensive working face [3,4,5], and the quantitative fault diagnosis and state degradation trend prediction of the key equipment [6,7]. In addition, it provides technical support for the guarantee of the safe and efficient production of the intelligent comprehensive working face and the reasonable maintenance of the equipment. Nevertheless, these systems are mainly aimed at researchers, and there is a lack of direct methods for general personnel or enterprise employees like the Q&A system. Each system of coal mining equipment is designed independently, and the problems of “information silos” and sub-system fragmentation exist among the sub-systems, which do not make appropriate management and direct use of the large amount of knowledge resources accumulated in the process of the intelligent production of coal mining equipment. Currently, many coal mining companies in China have gradually moved from traditional person-to-person investigation and control to computer technology-based information management in the management of coal mine equipment maintenance knowledge (CEMK). The fact is that there are still numerous coal mine enterprises that manage equipment maintenance and repair by reading a large number of maintenance manuals or consulting an expert. This leads to a restrictive and inefficient query. At present, there is no scientific, effective, and unified knowledge expression and structure composition in the field of knowledge in coal mine equipment maintenance. Structured knowledge is mostly stored in tables or forms, semi-structured knowledge is mostly stored in web pages related to coal mine equipment maintenance, and unstructured knowledge is often retained in the form of text in maintenance manuals, maintenance case bases, and maintenance standards. The correlations between the knowledge are not strong, and it is extremely difficult to obtain a unified form of the expression of knowledge. These problems have seriously hindered the efficiency and progress of the intelligent development and construction of coal mine equipment maintenance. Obviously, the degree of sharing and the reuse of multi-source coal mine equipment knowledge is low, so it is urgent and necessary to build an efficient knowledge management system, dig equipment maintenance knowledge for deep information, and enhance the level of equipment maintenance information.

3. Methodology

XCoalChat was proposed to train a professional large language model in the field of coal mine equipment maintenance by building a professional dataset for coal mine equipment maintenance. The performance of the model was improved through knowledge graph enhancement, the triple LoRA fine-tuning mechanism, and direct preference optimization so that it could better realize the three tasks of coal mine dialog consultation, professional consulting, and maintenance decision analysis. The specific framework for XCoalChat is illustrated in Figure 1.
Figure 1. A large language model for coal mine equipment maintenance in the XCoalChat framework.

3.1. Construction Dataset Based on ReliableCEMK-Self-Instruction

Comprehensively considering the two aspects of coal mine safety management and equipment maintenance, this study collated a large number of different coal mine related materials. We mainly referenced coal mine safety regulations, the fault knowledge corpus in cooperation with enterprises, and the operation instructions of different fully mechanized mining face equipment as long as it involved the fault maintenance decision records of the coal mine electrical equipment and the exam questions concerning the mining system, driving system, transport system, ventilation system, safety system, and coal processing system. This will contain the basic knowledge of the coal mining equipment, the professional operation knowledge, coal mining equipment intelligence, and other related contents investigated within the scope of this study. These texts mainly have structured, semi-structured, or unstructured multi-source representation, and this paper will adopt the corresponding processing means for different data types. For the structured data such as the operation data and staff information stored in the database of the equipment production monitoring system of coal mining enterprises, it can be transformed into the ternary format by means of direct mapping and defining rules; for semi-structured data such as data on the Internet and the web page data of the coal mine equipment forum, the data are extracted from the web page using a web wrapper and the extracted web page-like data into structured form; for unstructured data such as maintenance cases, integrity standards, and maintenance guide books of coal mine equipment, etc., the key information contained therein was obtained through natural language processing techniques.
ReliableCEMK-Self-Instruction, on the strength of specific knowledge in the field of mine equipment maintenance, enhances the reliability of the dataset. After providing correlatively specific knowledge text of the coal mine equipment, ChatGPT was first asked to generate several questions pertaining to the text and their logical relationship, and then ChatGPT was asked to answer the questions through the “text segment–question” pair. In this way, XCoalChat can generate answers containing relevant normative information to ensure the accuracy of the answers. Finally, a 92 k coal mine dialog dataset, 24 k coal mine examination dataset, and 32 k maintenance decision analysis dataset were constructed.

3.2. Knowledge Graph Enhancement Based on “Coal Mine Equipment Maintenance System—Full Life Cycle—Specification”

In an effort to enable LLM to identify information as well as the potential information from the coal mine equipment maintenance knowledge more systematically, completely and objectively, this paper constructed a three-dimensional coal mine equipment maintenance text database representation method based on the “Coal Mine Equipment System—Full Life Cycle—Specification”. As shown in Figure 2, the entire coal mine equipment system was divided into seven subsystems. Next, with the intention of highlighting the characteristics of problem evolution over time, the full life cycle phases of the coal mine equipment maintenance system were divided into the equipment installation stage, equipment operation stage, and equipment maintenance stage. Together with the division method of the coal mine equipment system, all problem text that needed to be identified in different periods and different subsystems were identified. Finally, through one-by-one analysis of the laws, regulations and technical specifications, industry standards, the literature, and the qualification certificate examination questions related to coal mine equipment maintenance, all of the text information that needed to be identified was classified. After integration, the potential information database of the coal mine equipment maintenance system was obtained. For example, “the shearer cannot be reversed due to damage of potentiometer or solenoid valve” was expressed in the form of “equipment operation phase—mining equipment system—maintenance record”.
Figure 2. 3D framework based on the “Coal mine Equipment Maintenance System—Full Life Cycle—Specification”.
Due to the lack of understanding of the deep meaning of the generated text in the large model, hallucinations occurred. For instance, when the LLM is asked how to cook reinforced concrete, the model may directly generate the recipe and the production process. This obviously contradicts our basic understanding. Knowledge graph enhancement improved the knowledge density, alleviating to some extent the illusion of large language models and the limitations of LLMs in performing tasks that require more knowledge about the world.
As shown in Figure 3, the knowledge graph enhancement method centered on the “Coal mine Equipment Maintenance System—Full Life Cycle—Specification” can guide the LLM to answer questions about downstream tasks. There is rich semantic knowledge among the knowledge graphs of coal mine equipment maintenance, that is, there are structural constraints between the triples.
Figure 3. Knowledge graph guidance problem.
As can be seen, if a pair of entities contains the types “failure phenomenon” and “failure cause”, the prediction probability will be relatively high in the relationship “failure phenomenon cause” and relatively low in the other relationships. Correspondingly, a relationship also constrains the type of its entity. An example of a knowledge graph enhancement prompt template, as shown in Table 1, converts each knowledge triplet into a prompt that allows the model to learn the constraints of a question pair and generate a response that contains both the constraint information and the target knowledge. As shown in Table 1, using the example Knowledge Graph Enhanced Prompt Template converts each knowledge triplet into a Prompt that allows the model to learn the constraint relationships of the Q&A pairs and generate a response that contains the constraint information with the target knowledge. When the input is “emulsion leakage”, the model initially identifies the hydraulic system according to the presence of “emulsion” and “leakage”. It then considers relevant constraints, infers components within the hydraulic system such as seals, and whether there are any defects or cracks. Finally, it generates a response indicating potential causes of failure.
Table 1. Example of a knowledge graph enhancement prompttemplate.

3.3. Large Language Model Construction Based on Triple-LoRA Fine-Tuning and DPO

3.3.1. Benchmark Model

In this work, the ChatGLM-6B model was chosen as the benchmark model. This is shown in Figure 4. ChatGLM-6B is an open-source conversational language model with support for Chinese-English bilingual Q&A and is optimized for Chinese. The model was founded on the general language model (GLM) architecture and has 6.2 billion parameters. The probability of text can be expressed as
p ( T ) = p ( w 1 ) p ( w 2 w 1 ) p ( w L w 1 , w 2 w L 1 )
where L stands for the size of the text.
Figure 4. Schematic diagram of the GLM.
The process is shown in Figure 5a,b. Concatenating the three fragments together gives the model input x 1 , x 2 , M , x 4 , M , S , x 5 , x 6 , S , and the model output is the occluded fragment, as shown in Figure 5c above. Two position codes are used here: the first code injects position information for the entire input, representing the location of the masked part in the original input, and the second code injects location information for tokens within the masked part. Position encoding uses rotary position encoding (RoPE) and it selects the GeLU activation function to optimize FFN. Training objectives include not only self-supervised GLM autoregressive blank filling, but also multi-task learning for a small number of tokens to improve the performance of their downstream zero-shot tasks.
Figure 5. Autoregressive blank infilling objective and self-attention masking. (a) The original text is sampled; (b) replace the sampled spans with M in Part A and shuffle the spans in Part B; (c) autoregressively generates Part B; (d) self-attention mask. The gray areas are obscured. Part A tokens can focus on themselves (blue frame), but not on Part B. Part B marking can focus on A and the antecedent in B (yellow and green frames correspond to two spans). M stands for [MASK], S stands for [START], E stands for [END].

3.3.2. Triple-LoRA Fine-Tuning Mechanisms

To ensure the specialization and task-oriented function of a coal mine equipment maintenance LLM, a multi-expert fine-tuning strategy was proposed. Low rank adaptive (LoRA), a mechanism for effective parameter fine-tuning [32], is used to train individual components of the model on a specialized dataset. They can operate autonomously without interfering with each other. LoRA applies adaptation to the pre-trained language model by introducing a low-rank matrix in its weight matrix. The pivotal steps in the LoRA adaptation process are as follows
  • Initialization: Start with a pre-trained language model and add a low-rank adaptation layer to its weight matrix array. This layer is represented by a low-rank matrix that is initialized to random values.
  • Fine-tuning: Train the model on a new task or domain, updating only the low-rank adaptation layer while keeping the weights of the pre-trained model fixed. This allows the model to learn task-specific information in a targeted way without changing its general knowledge.
  • Prediction: Use the fine-tuned model to make predictions about new, unseen data from the target task or domain.
LoRA indirectly trains some of the dense layers in a neural network. Afterward, LoRa employs trainable rank-decomposition matrices at each layer of the transformer architecture, thereby significantly reducing the required number of training parameters for downstream tasks. For the pre-trained weight matrix, LoRA uses low-rank decomposition, indicating the latter to constrain its updates.
W 0 + Δ W = W 0 + B A
where A d × k , B d × k and r min ( d , k ) . During training, W 0 is frozen and does not receive gradient updates, but A and B contain trainable parameters. Note that both W and Δ W = B A both multiply with the identical inputs, and their corresponding output vectors are added in the coordinate direction. The original forward propagation is h = W 0 x The modified forward propagation is:
h = W 0 x + Δ W x = W 0 x + B A x
Ultimately, the inference speed is increased and the inference delay is reduced.
Specifically, we trained the three LoRA modules with fine-tuning of the base model for the corresponding part of the instruction. Switching between different functions during deployment only requires replacing the LoRA parameters loaded on the current base model. This eliminates the need to retrain the entire model. In addition, this approach improves the efficiency of the model for each mission. In this paper, three modules were set: LoRA for coal mine equipment dialog consultation, LoRA for coal mine professional advisory service, and LoRA for maintenance program analysis and decision. The triple LoRA mechanism is illustrated in Figure 6.
Figure 6. Fine-tuning architecture for triple LoRA mechanism.
Coal mine equipment dialog consulting: The first LoRA model was designed to address the problems and consulting dialog challenges in the context of coal mine equipment including the basic knowledge of coal mine equipment such as the development of coal mine equipment and the introduction of fully mechanized coal mining equipment like related structures and components. It mainly provides basic dialog and consultation services.
Coal mine professional consulting services: The second LoRA model aimed to train the model in the standard operation of coal mine equipment such as the operation involved in the installation and operation of the equipment as well as the ability of professional skills consulting. The main object of this service are the operators who actually participate in coal mine production or carry out coal mine-related equipment.
Maintenance program analysis and decision: The third LoRA model was mainly used to deal with various tasks of coal mine equipment fault analysis. These tasks include, but are not limited to, fault diagnosis and maintenance decision-making in the mine text. On being asked about common failure topics such as overheating, overcurrent, gear failure etc., the model can be analyzed like an expert based on relevant documentation and eventually provide maintenance recommendations.

3.3.3. Direct Preference Data Optimization Model

The direct preference optimization model skips the explicit reward modeling step and avoids the process of reinforcement learning [33]. Starting from the loss of the RL model, the optimal solution of the reward maximization objective of the KL constraint is as follows:
π r ( y | x ) = 1 Z ( x ) π r e f ( y | x ) exp 1 β r x , y
where Z ( x ) = y π r e f ( y | x ) exp 1 β r x , y is a partition function. We can rearrange the formula to represent the reward function depending on the matching optimal strategy π r , reference strategy π r e f , and partition function Z
r x , y = β log π r ( y | x ) π r e f ( y | x ) + β log Z x
We developed a maximum likelihood target for the parameterized strategies. As this is similar to the reward modeling approach, the policy goal becomes:
L D P O = E log σ β log π θ ( y w | x ) π r e f ( y w | x ) β log π θ ( y l | x ) π r e f ( y l | x )
This converts the loss function of the reward function to the loss function of the policy function, skipping the explicit reward modeling step. The policy networks represent both language models and rewards. The σ is a sigmoid function and β is a hyperparameter, generally between 0.1 and 0.5. y w and y l denote the good response and the poor response in a given preference data, respectively. π θ ( y w | x ) represents the cumulative probability that the current policy model generates a good response given input x . π r e f represents the cumulative probability that the original reference model will generate a bad response given input x . The good/bad responses generated by the model were annotated with preferences using GPT-4 and then trained using DPO to achieve XCoalChat outputs that were more orientated toward human preferences.

4. Results and Discussion

To verify the feasibility and validity of the proposed model, this paper conducted experiments around the following four aspects and analyzed them in comparison with other mainstream LLMs.
Experiment 1: The experiment of dialog consultation for coal mine equipment was conducted on the dialog dataset of coal mine equipment with a training set of 73.6 k and a test set of 18.4 k. The purpose was to verify the ability of the model to generate answers.
Experiment 2: The experiment of professional consultant for coal mine equipment was conducted on the dataset of professional consultants for coal mine equipment with a training set of 19.2 k and a test set of 4.8 k. In the form of multiple-choice questions, the model was tested to determine its professional ability in coal mine equipment.
Experiment 3: The experiment of maintenance decision analysis used the maintenance decision analysis dataset with a training set of 25.6 k and a test set of 6.4 k to judge whether the model could generate corresponding maintenance decision schemes for equipment failures.
Experiment 4: The experiment of inference time comparison recorded and compared the inference time of the models during the above experiment.
For the purpose of validating the proposed multi-source text coal mine equipment maintenance model, comparison methods such as Bloom-7b and LLama-7b were employed. Bloom-7b is a large language model developed and open sourced by BigScience with seven billion references [34]. It is trained on a dataset containing multiple languages, allowing for smoother training on raw sequence length and better downstream performance. LLama-7b is an open and efficient large base language model published by Meta AI, based on LLama, which uses the efficient implementation of causal multi-head attention mechanisms to reduce memory usage and runtime, resulting in several “Alpaca family” models such as Vicuna and Alpaca [35].

4.1. Experimental Environment and Parameter Setting

This experiment was run under the deep learning server A100 with 40 GB of memory. The model proposed in this paper was based on a Pytorch1.10 deep learning framework and CUDA version 11.3.
Because of its open-source nature and ability to support multiple languages, ChatGLM was chosen as our benchmark large language model. Following the references with the actual referencing process, for the supervised fine-tuning process, we set the learning rate to 2 × 10−4, batch size to 128, maximum length to 1024, and trained across three epochs. For the DPO process, we applied a learning rate of 5 × 10−4 and a maximum length of 1024 to train 5000 steps. To effectively train the large language model, LoRA fine-tuned r to 8, alpha to 16, and dropout to 0.05.

4.2. Experimental Results Analysis

4.2.1. Coal Mine Equipment Dialog Consultation

An experiment in which prior historical dialogs were randomly used as input in sequence in light of evaluating the performance of dialog consultation for coal mining equipment was conducted. Sample responses are shown in Table 2. BLEU [36] and ROUGE [37] were used to assess the quality of the dialog. BLEU is a commonly used metric that compares a candidate translation to one or more reference texts based on n-element syntactic precision. The larger the value, the better. On the other hand, ROUGE is a particularly important index for evaluating automatic summarization and machine translation. This is useful as it focuses on the recall aspect of the generated abstracts by comparing them with the references. As with the former, the larger the value, the better.
Table 2. Example of coal mine equipment dialog.
The results of the experiment are illustrated in Table 3, showing that XCoalChat achieved excellent ratings on all metrics in the BLEU and ROUGE scores. It revealed that the dataset constructed in this paper contains more diverse and rich knowledge about coal mine equipment maintenance, and the knowledge density of the model was improved by the method of knowledge graph enhancement, which allows the model to cover the information provided by the reference text more effectively by generating responses under the consideration of the constraints between the knowledge triples.
Table 3. Comparison of the dialog consultation of coal mine equipment.

4.2.2. Coal Mine Equipment Professional Consulting

The categories evaluated in the coal mine equipment professional consulting comparison experiment were based on the classes with the highest frequency in the coal mining exam dataset. The categories selected included coal mining systems, boring systems, transport systems, coal preparation systems, ventilation systems, and safety management. The question and candidate options were given as inputs and the text answers generated by the model were subsequently used to calculate accuracy. A sample response is shown in Table 4.
Table 4. Example of coal mine equipment professional consulting.
The experimental results are shown in Figure 7, indicating that XCoalChat had an average accuracy of 77.21% and outperformed other LLMs in all evaluation categories. Specifically, XCoalChat achieved a strong performance, surpassing ChatGLM-6B, BLOOM-7B, and LLama7B’s average scores of 53.59%, 47.16%, and 48.42%, respectively. It was not surprising to find that the model output the referenced normative regulations when answering questions, which was due to the ReliableCEMK-Self-Instruction methodology used in this study, based on the specific knowledge in the field of coal mine equipment maintenance and greatly improved the reliability of the model. This makes the model’s answers more informative, and the model can provide coal mine equipment operators with a reasoned guide to equipment maintenance that can reduce the incidence of incorrect operation.
Figure 7. Experimental comparison of the professional skills of coal mine equipment. (a) Safety management system; (b) mining system; (c) ventilation system; (d) driving system; (e) coal processing system; (f) transport system.

4.2.3. Coal Mine Equipment Maintenance Decision

The maintenance decision analysis capability of the LLM was evaluated on the maintenance decision analysis dataset. Electrical faults, hydraulic faults, mechanical faults, and auxiliary system faults were selected for evaluation. With each fault record as the input and the generated text as the output, the accuracy of the models was calculated by comparing the generated text with the diagnostic labels in the fault records. Looking at Table 5, it is an example of a maintenance decision analysis response. This indicates that the knowledge graph enhancement guides the larger model to incorporate the underlying relationships between the fault phenomena, fault causes, and maintenance decisions in the reasoning process, which improves the model’s logical capability and interpretability. Providing DPO methods to correct the model makes the model more suitable to the field of coal mine equipment maintenance, generates more relevant responses, and the model has a strong ability to understand and explain different fault groups.
Table 5. Example of a coal mine equipment maintenance decision.
The experimental results for each failure group are displayed in Figure 8, where it can be seen that XCoalChat outperformed the other language models in all failure groups. XCoalChat had an average accuracy of 81.77% in all fault groups compared with ChatGLM-6B, LLama 7B, and Bloom-7B, which improved the accuracy by 47.96%, 37.25% and 39.69%, respectively. XCoalChat performed particularly strongly in terms of the hydraulic and mechanical failures, achieving an accuracy of 90.12% and 86.04%, respectively. This demonstrates the robustness of the model to understand and account for different groups of faults and will provide maintenance personnel with more accurate maintenance decision support. Maintenance specialists can refer to XCoalChat for a reference maintenance program, reducing the time spent.
Figure 8. Analysis and comparison of the coal mine equipment maintenance decision.

4.2.4. Inference Time Comparison

The reasoning time of the model in the above experiment was recorded, and the reasoning time of the base model was the benchmark for calculating the relative reasoning speeds of the other models and comparing their average time in each experiment; results are shown in Figure 9. Among the two tasks of dialog consultation and coal mine maintenance decision analysis, XCoalChat took the least reasoning time. The average time required for all task inference was the least, which was 89.66% of the running time of the benchmark model, and had a high practical efficiency. The triple LoRA fine-tuning mechanism enabled the model to maintain exceptional performance across all downstream tasks while significantly reducing the model’s computational overhead. XCoalChat still had a good reasoning speed while maintaining accuracy.
Figure 9. Inference time comparison.

5. Conclusions

Aiming at the characteristics of coal mine equipment maintenance knowledge such as complexity, dispersion, low sharing and a lack of effective management, a large language model, XCoalChat, in the coal mine equipment maintenance domain based on multi-source text, was proposed. The large language model was introduced as a pioneer into the field of coal mine equipment maintenance, and the vertical field dataset of coal mine equipment maintenance was especially prepared. Compared with other mainstream large models, the method in this paper had the highest scores of BLEU and ROUGE in the dialog consultation experiment of coal mine equipment; in the professional consulting experiment of coal mine equipment, the accuracy increased by 44.08%, 63.72%, and 59.46% compared with ChatGLM-6b, Bloom-7b, and LLaMA-7b, respectively. In the maintenance decision analysis experiment, the accuracy increased by 47.96%, 37.25%, and 39.69% compared with ChatGLM-6b, Bloom-7b, and LLaMA-7b, respectively. Furthermore, the entire model inference only requires 13.28 G of GPU, which can be deployed both personally and in enterprises.
In future research, due to the fact that there is also a large amount of complex and diverse multimodal coal mining equipment maintenance knowledge, we will conduct an in-depth study into the multi-modal large language model and multi-agent mechanism, combining document intelligence and the reasoning ability of the large model to constantly update and iterate as well as continue to promote mine intelligence. An exploration of potential interdisciplinary applications and social intelligence mines as a feature approach for such analysis will be considered in the future.

Author Contributions

Study conception and design: W.X. and X.C.; Data collection: W.X. and X.Y.; Analysis and interpretation of results: W.X. and Y.D.; Draft manuscript preparation: W.X. and J.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Natural Science Foundation of National Natural Science Foundation of China (Nos. 52274158 and 51834006).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data and materials used in this research are available upon reasonable request to the corresponding author. The data are not publicly available due to privacy.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Wang, G. New technological progress of coal mine intelligence and its problems. Coal Sci. Technol. 2022, 50, 1–27. [Google Scholar]
  2. Gao, Y.; Mu, J.; He, J.; Chen, H.; Mu, J.; An, Z.; Zhao, X.; Chou, A. Wireless Self-powered Condition Monitoring System for Coal Machine Equipment. J. Mech. Eng. 2020, 56, 41–49. [Google Scholar]
  3. Duan, Y.; Cao, X.; Zhao, J.; Zhang, R.; Yang, X.; Guo, X. A health indicator construction method of rolling bearing based on vibration image features and deep autoencoder network. In Proceedings of the 5th International Conference on System Reliability and Safety Engineering, Beijing, China, 20–23 October 2023. [Google Scholar]
  4. Wang, Z.; Xu, X.; Si, L.; Ji, R.; Liu, X.; Tan, C. A Dynamic Health Assessment Approach for Shearer Based on Artificial Immune Algorithm. Comput. Intell. Neurosci. 2016, 2016, 9674942. [Google Scholar] [CrossRef] [PubMed]
  5. Si, L.; Wang, Z.; Liu, Z.; Liu, X.; Tan, C.; Xu, R. Health condition evaluation for a shearer through the integration of a fuzzy neural network and improved particle swarm optimization algorithm. Appl. Sci. 2016, 6, 171. [Google Scholar] [CrossRef]
  6. Zhang, W.; Li, J.; Wu, L.; Li, B. Research on fault diagnosis of idler bearing of belt conveyor based on 1DCNN-ELM. J. Coal Sci. Technol. 2023, 51, 383–389. [Google Scholar]
  7. Li, X.; Geng, Y.; Wang, H.; Fu, X.; Wang, R. Prediction method for residual life of shearer bearings based on improved similarity model. J. Mine Autom. 2022, 49, 96–103. [Google Scholar]
  8. Wang, Y.; Li, Q.; Chang, M.; Chen, H.; Zang, G. Research on Fault Diagnosis Expert System Based on the Neural Network and the Fault Tree Technology. Proced. Eng. 2012, 31, 1206–1210. [Google Scholar] [CrossRef]
  9. Pan, L.; Zhang, J.; Zhang, Y.; Xie, J. Comput. Appl. Softw. 2019, 36, 47–64.
  10. Li, Z.; Zhou, B.; Li, W.; Li, X.; Zhou, Y.; Feng, Z.; Zhao, H. Construction and application of mine electromechanical equipment accident knowledge graph. J. Mine Autom. 2022, 48, 107–110. [Google Scholar]
  11. Cao, X.; Zhang, M.; Lei, Z.; Duan, X.; Chen, R. Construction and application of knowledge graph for coal mine equipment maintenance. J. Mine Autom. 2021, 47, 41–45. [Google Scholar]
  12. Ye, S. Research on the Construction and Query Method of Knowledge Graph in Coalmine Based on Neo4j. Master’s Thesis, China University of Mining and Technology, Xuzhou, China, 2019. [Google Scholar]
  13. Liu, P.; Ye, S.; Shu, Y.; Lu, X.; Liu, M. Coalmine Safety: Knowledge Graph Construction and Its QA Approach. J. Chin. Inf. Process. 2020, 34, 49–59. [Google Scholar]
  14. Zhang, D.; Liu, Z.; Jia, W.; Liu, H.; Tan, J. A Review on Knowledge Graph and Its Application Prospects to Intelligent Manufacturing. J. Mech. Eng. 2021, 57, 90–113. [Google Scholar]
  15. Patil, R.; Gudivada, V. A Review of Current Trends, Techniques, and Challenges in Large Language Models (LLMs). Appl. Sci. 2024, 14, 2074. [Google Scholar] [CrossRef]
  16. Zhao, W.; Zhou, K.; Li, J.; Tang, T.; Wang, X.; Hou, Y.; Min, Y.; Zhang, B.; Zhang, J.; Dong, Z.; et al. A Survey of Large Language Models. arXiv 2023, arXiv:2303.18223. [Google Scholar]
  17. Ouyang, L.; Wu, J.; Jiang, X.; Almeida, D.; Wainwright, C.L.; Mishkin, P.; Zhang, C.; Agarwal, S.; Slama, K.; Ray, A.; et al. Training language models to follow instructions with human feedback. arXiv 2022, arXiv:2203.02155. [Google Scholar]
  18. OpenAI. Gpt-4 Technical Report. 2023. arXiv. 2023, arXiv:2303.08774. [Google Scholar]
  19. Chowdhery, A.; Narang, S.; Devlin, J.; Bosma, M.; Mishra, G.; Roberts, A.; Barham, P.; Chung, H.W.; Sutton, C.; Gehrmann, C.; et al. PaLM: Scaling Language Modeling with Pathways. J. Mach. Learn. Res. 2022, 24, 240:1–240:113. [Google Scholar]
  20. Zeng, A.; Liu, X.; Du, Z.; Wang, Z.; Lai, H.; Ding, M.; Yang, Z.; Xu, Y.; Zheng, W.; Xia, X.; et al. GLM-130B: An Open Bilingual Pre-trained Model. arXiv 2022, arXiv:2210.02414. [Google Scholar]
  21. Trad, F.; Chehab, A. Prompt Engineering or Fine-Tuning? A Case Study on Phishing Detection with Large Language Models. Mach. Learn. Knowl. Extr. 2024, 6, 367–384. [Google Scholar] [CrossRef]
  22. Liu, X.; Wang, G.; Zha, D. FinGPT: Democratizing Internet-scale Data for Financial Large Language Models. arXiv 2023, arXiv:2307.10485. [Google Scholar]
  23. Zhang, X.; Yang, Q.; Xu, D. XuanYuan 2.0: A Large Chinese Financial Chat Model with Hundreds of Billions. Parameters. In Proceedings of the 32nd ACM International Conference on Information and Knowledge Management, Birmingham, UK, 21–25 October 2023. [Google Scholar]
  24. Cui, J.; Li, Z.; Yan, Y.; Chen, B.; Yuan, L. ChatLaw: Open-Source Legal Large Language Model with Integrated External Knowledge Bases. arXiv 2023, arXiv:2306.16092. [Google Scholar]
  25. Huang, Q.; Tao, M.; An, Z.; Zhang, C.; Jiang, C.; Chen, Z.; Wu, Z.; Feng, Y. Lawyer LLaMA Technical Report. arXiv 2023, arXiv:2305.15062. [Google Scholar]
  26. Wang, H.; Liu, C.; Xi, N.; Qiang, Z.; Zhao, S.; Qin, B.; Liu, T. HuaTuo: Tuning LLaMA Model with Chinese Medical Knowledge. arXiv 2023, arXiv:2304.06975. [Google Scholar]
  27. Li, Y.; Li, Z.; Zhang, K.; Dan, R.; Zhang, Y. ChatDoctor: A Medical Chat Model Fine-tuned on LLaMA Model using Medical Domain Knowledge. arXiv 2023, arXiv:2303.14070. [Google Scholar]
  28. Xiong, H.; Wang, S.; Zhu, Y.; Zhao, Z.; Liu, Y.; Huang, L.; Wang, Q.; Shen, D. DoctorGLM: Fine-tuning your Chinese Doctor is not a Herculean Task. arXiv 2023, arXiv:2304.01097. [Google Scholar]
  29. Zheng, O.; Abdel-Aty, M.A.; Wang, D.; Wang, C.; Ding, S. TrafficSafetyGPT: Tuning a Pre-trained Large Language Model to a Domain-Specific Expert in Transportation Safety. arXiv 2023, arXiv:2307.15311. [Google Scholar]
  30. Lowin, M. A Text-Based Predictive Maintenance Approach for Facility Management Requests Utilizing Association Rule Mining and Large Language Models. Mach. Learn. Knowl. Extr. 2024, 6, 233–258. [Google Scholar] [CrossRef]
  31. Wang, Y.; Kordi, Y.; Mishra, S.; Liu, A.; Smith, N.A.; Khashabi, D.; Hajishirzi, H. Self-Instruct: Aligning Language Models with Self-Generated Instructions. Annual Meeting of the Association for Computational Linguistics. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Toronto, ON, Canada, 9–14 July 2023. [Google Scholar]
  32. Hu, J.E.; Shen, Y.; Wallis, P.; Allen-Zhu, Z.; Li, Y.; Wang, S.; Chen, W. LoRA: Low-Rank Adaptation of Large Language Models. arXiv 2021, arXiv:2106.09685. [Google Scholar]
  33. Rafailov, R.; Sharma, A.; Mitchell, E.; Ermon, S.; Manning, C.D.; Finn, C. Direct Preference Optimization: Your Language Model is Secretly a Reward Model. arXiv 2023, arXiv:2305.18290. [Google Scholar]
  34. Le Scao, T.; Fan, A.; Akiki, C.; Pavlick, E.; Ilić, S.; Hesslow, D.; Castagné, R.; Luccioni, A.S.; Yvon, F.; Gallé, M.; et al. Bloom: A 176b-parameter open-access multilingual language model. arXiv 2022, arXiv:2211.05100. [Google Scholar]
  35. Touvron, H.; Lavril, T.; Izacard, G.; Martinet, X.; Lachaux, M.; Lacroix, T.; Rozière, B.; Goyal, N.; Hambro, E.; Azhar, F.; et al. LLaMA: Open and Efficient Foundation Language Models. arXiv 2023, arXiv:2302.13971. [Google Scholar]
  36. Papineni, K.; Roukos, S.; Ward, T.; Zhu, W. Bleu: A Method for Automatic Evaluation of Machine Translation. In Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics, Philadelphia, PA, USA, 7–12 July 2002. [Google Scholar]
  37. Lin, C. ROUGE: A Package for Automatic Evaluation of Summaries. In Proceedings of the 42nd Annual Meeting on Association for Computational Linguistics, Barcelona, Spain, 21–26 July 2004. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.