A Fault Diagnosis Framework for Waterjet Propulsion Pump Based on Supervised Autoencoder and Large Language Model
Abstract
1. Introduction
- Firstly, we abandoned the traditional LLM fault diagnosis framework of manually extracting time–frequency domain features. Instead, we employ a supervised autoencoder (SAE) to directly learn deep features from raw vibration signals that are oriented toward the fault classification task. This achieves automated feature extraction while ensuring that the features optimally discriminate between different fault types.
- Secondly, a deep, integrated, intelligent diagnostic framework was constructed, which combines the high-quality features learned by the SAE with the powerful generalization and reasoning capabilities of an LLM. By using parameter-efficient fine-tuning techniques, we construct an end-to-end, high-precision diagnostic model.
- Finally, we comprehensively validated the superiority of the proposed framework by conducting comprehensive experiments under two key scenarios: direct diagnosis with full data and transfer diagnosis under limited-data conditions. The results showed that compared to the baseline method using manually extracted time–frequency features, our framework significantly improves diagnostic accuracy, generalization, and data efficiency.
2. Related Work
2.1. Based on Data-Driven Fault Diagnosis Techniques
2.2. Text Sequence Format Input in LLM-Based Fault Diagnosis
2.3. LLM Fine-Tuning Technology in Fault Diagnosis
2.4. Gap Analysis
- Suboptimal LLM inputs: Current LLM-based diagnosis methods either rely on suboptimal, time-consuming manual extraction of time–frequency features, or on chunking raw signals directly (which may fail to effectively abstract key fault information). The quality of the input representation fundamentally limits the LLM’s performance.
- Integrated diagnostic framework combining automated feature learning and LLMs: Automated feature learning methods and LLMs each have advantages in classification generalization, but a collaborative framework that unifies the two for fault diagnosis remains to be explored. In particular, using features explicitly optimized for classification tasks as inputs to LLMs is a highly promising yet underexplored direction.
3. Methodology
3.1. Overall Framework
- Baseline method: The baseline method serves as the experimental benchmark. This approach follows the idea proposed by Tao et al. [16]. It aims to evaluate the diagnostic performance when combining mature, traditional signal processing techniques with cutting-edge LLMs. Its core idea is to leverage expert knowledge to manually extract 24 validated statistical features from the raw signal (in both time and frequency domains) and then convert these structured numerical features into a natural language format that an LLM can understand. In this work, we select a total of 20 features (time and frequency domain), having removed several redundant features.
- Proposed SAE-LLM method: This approach directly inputs raw vibration signal segments into a pre-trained SAE encoder, which automatically extracts a 20-dimensional deep feature vector. This feature vector is then fed into an LLM for fine-tuning to perform fault classification.
3.2. Proposed Method: LLM Diagnosis Framework Based on Supervised Autoencoder
3.2.1. Method Overview
3.2.2. Framework Composition
- SAE model training: Design and train an SAE model. This model takes raw signal segments as input and optimizes both the reconstruction and classification objectives simultaneously.
- Automated feature extraction: Use the trained SAE encoder to directly map new vibration signal segments to a 20-dimensional deep feature vector.
- Feature textualization and LLM fine-tuning: Following the same approach as the baseline method, textualize the feature vector extracted by the SAE and use it to fine-tune the LLM.
3.2.3. Structure of the Supervised Autoencoder (SAE)
3.3. LLM Fine-Tuning Technique
4. Experiments and Analysis
4.1. Experimental Platform and Data Collection
4.2. Experimental Parameter Settings
4.3. Experiment 1: Full Data Direct Diagnosis Performance Analysis
4.4. Experiment 2: Limited Data Transfer Diagnosis Performance Analysis
- Scenario a: Fine-tune the pre-trained LLM using all the data from conditions A and B, as well as 10% of the data from condition C, then test on the remaining 90% of condition C.
- Scenario b: Fine-tune the pre-trained LLM using only 10% of the data from condition C as the training set, then test on the remaining 90% of condition C.
5. Conclusions
- Automated, high-accuracy diagnostic paradigm: We successfully constructed and validated a new automated diagnostic paradigm. Experimental results show that, in both data-rich direct diagnosis scenarios and limited-data, cross-condition transfer scenarios, the proposed SAE-LLM framework significantly outperforms the baseline method based on manual features in terms of diagnostic accuracy.
- Critical role of task-oriented feature learning: We demonstrated that learning task-specific features is crucial for the LLM’s diagnostic performance. Through the SAE’s supervised learning mechanism, the extracted features are learned automatically, and more importantly, their inherent fault discriminability far exceeds that of manual statistical features. This allows the LLM to fully exploit its classification potential, especially in distinguishing confusing compound faults.
- Strong generalization under practical industrial conditions: The excellent performance of the proposed framework in limited data and transfer learning experiments demonstrates its strong generalization ability and data efficiency. This provides an effective technical solution for addressing the common issues of data scarcity and varying operating conditions in intelligent industrial maintenance.
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Molland, A.F.; Turnock, S.R.; Hudson, D.A. Ship Resistance and Propulsion; Cambridge University Press: Cambridge, UK, 2017. [Google Scholar]
- Gong, J.; Guo, C.; Wang, C.; Wu, T.; Song, K. Analysis of waterjet-hull interaction and its impact on the propulsion performance of a four-waterjet-propelled ship. Ocean. Eng. 2019, 180, 211–222. [Google Scholar] [CrossRef]
- Li, G.; Geng, H.; Xie, F.; Xu, C.; Xu, Z. Ensemble Deep Transfer Learning Method for Fault Diagnosis of Waterjet Pump Under Variable Working Conditions. Ship Boat 2025, 36, 103. [Google Scholar]
- Lei, Y.; Yang, B.; Jiang, X.; Jia, F.; Li, N.; Nandi, A.K. Applications of machine learning to machine fault diagnosis: A review and roadmap. Mech. Syst. Signal Process. 2020, 138, 106587. [Google Scholar] [CrossRef]
- Randall, R.B.; Antoni, J. Rolling element bearing diagnostics—A tutorial. Mech. Syst. Signal Process. 2011, 25, 485–520. [Google Scholar] [CrossRef]
- Zhang, R.; Tao, H.; Wu, L.; Guan, Y. Transfer learning with neural networks for bearing fault diagnosis in changing working conditions. IEEE Access 2017, 5, 14347–14357. [Google Scholar] [CrossRef]
- Hoang, D.-T.; Kang, H.-J. A survey on deep learning based bearing fault diagnosis. Neurocomputing 2019, 335, 327–335. [Google Scholar] [CrossRef]
- Lv, F.; Wen, C.; Bao, Z.; Liu, M. Fault diagnosis based on deep learning. In Proceedings of the 2016 American Control Conference (ACC), Boston, MA, USA, 6–8 July 2016; pp. 6851–6856. [Google Scholar]
- Zhang, W.; Li, C.; Peng, G.; Chen, Y.; Zhang, Z. A deep convolutional neural network with new training methods for bearing fault diagnosis under noisy environment and different working load. Mech. Syst. Signal Process. 2018, 100, 439–453. [Google Scholar] [CrossRef]
- Pan, H.; He, X.; Tang, S.; Meng, F. An improved bearing fault diagnosis method using one-dimensional CNN and LSTM. Stroj. Vestn. J. Mech. Eng. 2018, 64, 443–453. [Google Scholar]
- Chen, X.; Yang, R.; Xue, Y.; Huang, M.; Ferrero, R.; Wang, Z. Deep transfer learning for bearing fault diagnosis: A systematic review since 2016. IEEE Trans. Instrum. Meas. 2023, 72, 1–21. [Google Scholar] [CrossRef]
- Ding, Y.; Ma, L.; Ma, J.; Wang, C.; Lu, C. A generative adversarial network-based intelligent fault diagnosis method for rotating machinery under small sample size conditions. IEEE Access 2019, 7, 149736–149749. [Google Scholar] [CrossRef]
- Brown, T.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.D.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A. Language models are few-shot learners. Adv. Neural Inf. Process. Syst. 2020, 33, 1877–1901. [Google Scholar]
- Chang, Y.; Wang, X.; Wang, J.; Wu, Y.; Yang, L.; Zhu, K.; Chen, H.; Yi, X.; Wang, C.; Wang, Y. A survey on evaluation of large language models. ACM Trans. Intell. Syst. Technol. 2024, 15, 1–45. [Google Scholar] [CrossRef]
- Naveed, H.; Khan, A.U.; Qiu, S.; Saqib, M.; Anwar, S.; Usman, M.; Akhtar, N.; Barnes, N.; Mian, A. A comprehensive overview of large language models. ACM Trans. Intell. Syst. Technol. 2025. just accepted. [Google Scholar] [CrossRef]
- Tao, L.; Liu, H.; Ning, G.; Cao, W.; Huang, B.; Lu, C. LLM-based framework for bearing fault diagnosis. Mech. Syst. Signal Process. 2025, 224, 112127. [Google Scholar] [CrossRef]
- Peng, Z.K.; Chu, F. Application of the wavelet transform in machine condition monitoring and fault diagnostics: A review with bibliography. Mech. Syst. Signal Process. 2004, 18, 199–221. [Google Scholar] [CrossRef]
- Zio, E. Prognostics and Health Management (PHM): Where are we and where do we (need to) go in theory and practice. Reliab. Eng. Syst. Saf. 2022, 218, 108119. [Google Scholar] [CrossRef]
- Chen, Z.; He, G.; Li, J.; Liao, Y.; Gryllias, K.; Li, W. Domain adversarial transfer network for cross-domain fault diagnosis of rotary machinery. IEEE Trans. Instrum. Meas. 2020, 69, 8702–8712. [Google Scholar] [CrossRef]
- Li, F.; Tang, T.; Tang, B.; He, Q. Deep convolution domain-adversarial transfer learning for fault diagnosis of rolling bearings. Measurement 2021, 169, 108339. [Google Scholar] [CrossRef]
- Li, T.; Zhao, Z.; Sun, C.; Yan, R.; Chen, X. Domain adversarial graph convolutional network for fault diagnosis under variable working conditions. IEEE Trans. Instrum. Meas. 2021, 70, 3515010. [Google Scholar] [CrossRef]
- Li, Y.; Song, Y.; Jia, L.; Gao, S.; Li, Q.; Qiu, M. Intelligent fault diagnosis by fusing domain adversarial training and maximum mean discrepancy via ensemble learning. IEEE Trans. Ind. Inform. 2020, 17, 2833–2841. [Google Scholar] [CrossRef]
- Wen, L.; Gao, L.; Li, X. A new deep transfer learning based on sparse auto-encoder for fault diagnosis. IEEE Trans. Syst. Man Cybern. Syst. 2017, 49, 136–144. [Google Scholar] [CrossRef]
- Li, Y.-F.; Wang, H.; Sun, M. ChatGPT-like large-scale foundation models for prognostics and health management: A survey and roadmaps. Reliab. Eng. Syst. Saf. 2024, 243, 109850. [Google Scholar] [CrossRef]
- Tao, L.; Li, S.; Liu, H.; Huang, Q.; Ma, L.; Ning, G.; Chen, Y.; Wu, Y.; Li, B.; Zhang, W. An outline of Prognostics and health management Large Model: Concepts, Paradigms, and challenges. Mech. Syst. Signal Process. 2025, 232, 112683. [Google Scholar] [CrossRef]
- Hinton, G.E.; Zemel, R. Autoencoders, minimum description length and Helmholtz free energy. Adv. Neural Inf. Process. Syst. 1993, 6, 3–10. [Google Scholar]
- Lu, C.; Wang, Z.-Y.; Qin, W.-L.; Ma, J. Fault diagnosis of rotary machinery components using a stacked denoising autoencoder-based health state identification. Signal Process. 2017, 130, 377–388. [Google Scholar] [CrossRef]
- Sun, M.; Wang, H.; Liu, P.; Huang, S.; Fan, P. A sparse stacked denoising autoencoder with optimized transfer learning applied to the fault diagnosis of rolling bearings. Measurement 2019, 146, 305–314. [Google Scholar] [CrossRef]
- Ding, N.; Qin, Y.; Yang, G.; Wei, F.; Yang, Z.; Su, Y.; Hu, S.; Chen, Y.; Chan, C.-M.; Chen, W. Parameter-efficient fine-tuning of large-scale pre-trained language models. Nat. Mach. Intell. 2023, 5, 220–235. [Google Scholar] [CrossRef]
- He, J.; Zhou, C.; Ma, X.; Berg-Kirkpatrick, T.; Neubig, G. Towards a unified view of parameter-efficient transfer learning. arXiv 2021, arXiv:2110.04366. [Google Scholar]
- Han, Z.; Gao, C.; Liu, J.; Zhang, J.; Zhang, S.Q. Parameter-efficient fine-tuning for large models: A comprehensive survey. arXiv 2024, arXiv:2403.14608. [Google Scholar]
- Li, X.L.; Liang, P. Prefix-tuning: Optimizing continuous prompts for generation. arXiv 2021, arXiv:2101.00190. [Google Scholar]
- Houlsby, N.; Giurgiu, A.; Jastrzebski, S.; Morrone, B.; De Laroussilhe, Q.; Gesmundo, A.; Attariyan, M.; Gelly, S. Parameter-efficient transfer learning for NLP. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; pp. 2790–2799. [Google Scholar]
- Lester, B.; Al-Rfou, R.; Constant, N. The power of scale for parameter-efficient prompt tuning. arXiv 2021, arXiv:2104.08691. [Google Scholar]
- Hu, E.J.; Shen, Y.; Wallis, P.; Allen-Zhu, Z.; Li, Y.; Wang, S.; Wang, L.; Chen, W. Lora: Low-rank adaptation of large language models. ICLR 2022, 1, 3. [Google Scholar]
- Chen, Y.; Liu, C. Sequential multi-objective multi-agent reinforcement learning approach for system predictive maintenance of turbofan engine. Adv. Eng. Inform. 2025, 67, 103553. [Google Scholar] [CrossRef]
Working Condition | Fault No. | Fault Mode | Spray Pump Speed (rpm) | Number of Samples |
---|---|---|---|---|
0 | 1 | Normal | 100 | 150 |
2 | Bearing Fault | 150 | ||
3 | Compound Fault | 150 | ||
1 | 1 | Normal | 150 | 190 |
2 | Bearing Fault | 190 | ||
3 | Compound Fault | 190 | ||
2 | 1 | Normal | 200 | 190 |
2 | Bearing Fault | 190 | ||
3 | Compound Fault | 190 |
Model | Hyperparameters | Value |
---|---|---|
Supervised Autoencoder | optimizer | Adam |
learning rate | 10−3 | |
batch size | 64 | |
training epoch | 100 | |
loss weight λ | 0.5 | |
LLM fine-tuning | fine-tuning techniques | LoRA |
lora_rank | 8 | |
lora_alpha | 16 | |
learning rate | 5 × 10−5 | |
per_device_train_batch_size | 2 | |
gradient_accumulation_steps | 8 |
Components | Layers | Parameters |
---|---|---|
Encoder | Linear | 3000 × 512 + Relu |
BatchNorm1d | 512 × 512 | |
Dropout | 512 × 512, dropout rate: 0.2 | |
Linear | 512 × 256 + Relu | |
BatchNorm1d | 256 × 256 | |
Dropout | 256 × 256, dropout rate: 0.2 | |
Linear | 256 × 128 + Relu | |
BatchNorm1d | 128 × 128 | |
Linear | 128 × 20 | |
Decoder | Linear | 20 × 128 + Relu |
BatchNorm1d | 128 × 128 | |
Dropout | 128 × 128, dropout rate: 0.2 | |
Linear | 128 × 256 + Relu | |
BatchNorm1d | 256 × 256 | |
Dropout | 256 × 256, dropout rate: 0.2 | |
Linear | 256 × 512 + Relu | |
BatchNorm1d | 512 × 512 | |
Linear | 512 × 3000 | |
Classifier | Linear | 20 × 64 + Relu |
BatchNorm1d | 64 × 64 | |
Dropout | 64 × 64, dropout rate: 0.2 | |
Linear | 64 × 32 + Relu | |
Dropout | 32 × 32, dropout rate: 0.2 | |
Linear | 32 × 3 |
Epoch | Llama-3.2-3B-Instruct | Qwen-2.5-3B-Instruct | Gemma-3-4B-Instruct |
---|---|---|---|
Accuracy | |||
1 | 63.52% | 66.67% | 78.62% |
2 | 90.25% | 67.30% | 94.34% |
3 | 91.19% | 65.09% | 99.06% |
4 | 93.08% | 66.67% | 98.43% |
5 | 91.82% | 66.67% | 99.06% |
6 | 92.14% | 89.62% | 99.69% |
7 | 91.51% | 92.45% | 99.69% |
8 | 93.08% | 99.06% | 99.69% |
9 | 93.08% | 92.77% | 99.06% |
10 | 93.40% | 98.43% | 99.37% |
Epoch | Llama-3.2-3B-Instruct | Qwen-2.5-3B-Instruct | Gemma-3-4B-Instruct |
---|---|---|---|
Accuracy | |||
1 | 97.17% | 98.74% | 78.62% |
2 | 97.17% | 99.37% | 94.34% |
3 | 97.80% | 99.69% | 99.06% |
4 | 97.48% | 99.69% | 98.43% |
5 | 97.48% | 100.00% | 100.00% |
6 | 98.11% | 100.00% | 100.00% |
7 | 97.80% | 100.00% | 100.00% |
8 | 97.80% | 99.69% | 100.00% |
9 | 98.11% | 100.00% | 99.69% |
10 | 98.43% | 100.00% | 100.00% |
Experimental No. | Train Condition | Test Condition | Accuracy (Epoch = 10) |
---|---|---|---|
a | Condition 0.1 + 10% Condition 2 | Condition 2 | 91.62% |
Condition 0.2 + 10% Condition 1 | Condition 1 | 62.77% | |
Condition 1.2 + 10% Condition 0 | Condition 0 | 62.96% | |
b | 10% Condition 2 | Condition 2 | 65.89% |
10% Condition 1 | Condition 1 | 34.70% | |
10% Condition 0 | Condition 0 | 40.99% |
Experimental No. | Train Condition | Test Condition | Accuracy (Epoch = 10) |
---|---|---|---|
a | Condition 0.1 + 10% Condition 2 | Condition 2 | 99.81% |
Condition 0.2 + 10% Condition 1 | Condition 1 | 100.00% | |
Condition 1.2 + 10% Condition 0 | Condition 0 | 98.02% | |
b | 10% Condition 2 | Condition 2 | 97.08% |
10% Condition 1 | Condition 1 | 83.43% | |
10% Condition 0 | Condition 0 | 78.77% |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Liu, Z.; Xiao, H.; Zhang, T.; Li, G. A Fault Diagnosis Framework for Waterjet Propulsion Pump Based on Supervised Autoencoder and Large Language Model. Machines 2025, 13, 698. https://doi.org/10.3390/machines13080698
Liu Z, Xiao H, Zhang T, Li G. A Fault Diagnosis Framework for Waterjet Propulsion Pump Based on Supervised Autoencoder and Large Language Model. Machines. 2025; 13(8):698. https://doi.org/10.3390/machines13080698
Chicago/Turabian StyleLiu, Zhihao, Haisong Xiao, Tong Zhang, and Gangqiang Li. 2025. "A Fault Diagnosis Framework for Waterjet Propulsion Pump Based on Supervised Autoencoder and Large Language Model" Machines 13, no. 8: 698. https://doi.org/10.3390/machines13080698
APA StyleLiu, Z., Xiao, H., Zhang, T., & Li, G. (2025). A Fault Diagnosis Framework for Waterjet Propulsion Pump Based on Supervised Autoencoder and Large Language Model. Machines, 13(8), 698. https://doi.org/10.3390/machines13080698