Domain-Specific Knowledge Graph for Quality Engineering of Continuous Casting: Joint Extraction-Based Construction and Adversarial Training Enhanced Alignment
Abstract
:1. Introduction
- A joint knowledge extraction method for quality engineering of continuous casting based on self-attention partition and recombination.
- A knowledge alignment method for continuous casting based on semi-supervised incremental learning and adversarial training.
2. Related Works
2.1. Knowledge Extraction
2.2. Knowledge Alignment
2.3. Domain-Specific Knowledge Graph
3. Problem Modeling
3.1. Characteristics of Quality Engineering Texts of Continuous Casting
3.2. DSKG Modeling for Quality Engineering of Continuous Casting
3.2.1. Constructing of Abstract KG
3.2.2. Constructing of Concrete KG
4. Method
4.1. Joint Extraction Based on Self-Attention Partition and Recombination for Quality Engineering of Continuous Casting
Algorithm 1 SAPRM |
Input: Continuous casting quality sequence T Output: Continuous casting quality triple set <entity, relation, entity> for each t in T do ρs = ρs + (t*e ∩ t*r) //get shared partition from token t with entity gate and relation gate ρe = ρe + (t*e − ρs) //get NER-related partition from token t with entity gate and shared partition ρr = ρs + (t*r − ρs) //get RE-related partition from token t with relation gate and shared partition oe, os, or = attention (ρe, ρs, ρr) //calculate attention score in attention module μe = oe + os, μr = or + os he = tanh(μe), hr = tanh(μr) for pos in length(T) do if is_head_start_pos(pos, he) is true then entities += (T[pos, end_pos], k) for e1, e2 in entities do if is_correct_relation (e1, e2, hr) is true then triples += (e1, r, e2) return continuous casting quality triple set |
4.1.1. Partition and Recombination Module of Continuous Casting Quality Sequence
- Partition of Continuous Casting Quality Sequence
- Self-Attention Mechanism
- Recombination of Continuous Casting Quality Sequence
4.1.2. NER Module of Continuous Casting Quality
4.1.3. RE Module of Continuous Casting Quality
4.1.4. Loss Function
4.2. Knowledge Alignment Method for Quality Engineering of Continuous Casting Based on Semi-Supervised Incremental Learning
Algorithm 2 Knowledge Alignment Method Based on Semi-supervised Incremental Learning. |
Input: labeled pairs of entities Wlabel and unlabeled pairs of entities Wunlabel Output: finely-tuned BERT BERTFT ← fine-tune(BERT(e1, e2)) for (e1, e2) in Wlabel for (e1, e2) in Wunlabel pl ← BERTFT((e1, e2)) Wtotal ← Wtotal + (e1, e2) if confidence(pl) ≥ threshold for w in Wunlabel ep ← replace w with w’ if importance(w) ≥ threshold m ← BERTFT(ep, e2) Wtotal ← Wtotal + (ep, e2) if m changed BERTFT ← fine-tune(BERT(e1, e2)) for (e1, e2) in Wtotal return BERTFT |
4.2.1. Supervised Learning Module
4.2.2. Pseudo-Label Generation Module for Continuous Casting Quality Knowledge
4.2.3. Incremental Learning Module of Continuous Casting Quality Knowledge
5. Experiment and Analysis
5.1. Experimental Environment and Parameters Setting
5.1.1. Experimental Data
5.1.2. Hyperparameter Setting of DSKG Construction Method for Quality Engineering of Continuous Casting
5.2. Evaluation of SAPRM
5.2.1. Performance of SAPRM
5.2.2. Ablation Study on SAPRM
5.2.3. Experiment on Loss Function
5.2.4. Generalization Capability
5.3. Evaluation of Continuous Casting Quality Knowledge Alignment Method
5.3.1. Performance of Continuous Casting Quality Knowledge Alignment Method
5.3.2. Ablation Study on Continuous Casting Quality Knowledge Alignment Method
5.4. Storage of DSKG for Quality Engineering of Continuous Casting
6. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Birs, I.R.; Muresan, C.; Copot, D.; Ionescu, C.-M. Model identification and control of electromagnetic actuation in continuous casting process with improved quality. IEEE-CAA J. Autom. Sin. 2023, 10, 203–215. [Google Scholar] [CrossRef]
- Thomas, B.G. Review on modeling and simulation of continuous casting. Steel Res. Int. 2018, 89, 1700312. [Google Scholar] [CrossRef]
- Abu-Salih, B. Domain-specific knowledge graphs: A survey. J. Netw. Comput. Appl. 2021, 185, 103076. [Google Scholar] [CrossRef]
- Li, Y. Research on Some Problems of Power Equipment Defect Text Mining and Application Based on Knowledge Graph. Master’s Thesis, Zhejiang University, Hangzhou, China, 2022. [Google Scholar]
- Chandak, P.; Huang, K.; Zitnik, M. Building a knowledge graph to enable precision medicine. Sci. Data 2023, 10, 67. [Google Scholar] [CrossRef] [PubMed]
- Yang, Z.; Huang, Y.; Feng, J. Learning to leverage high-order medical knowledge graph for joint entity and relation extraction. In Findings of the Association for Computational Linguistics: ACL 2023; Association for Computational Linguistics: Toronto, ON, Canada, 2023; pp. 9023–9035. [Google Scholar]
- Sun, Z.; Xing, L.; Zhang, L.; Cai, H.; Guo, M. Joint biomedical entity and relation extraction based on feature filter table labeling. IEEE Access 2023, 11, 127422–127430. [Google Scholar] [CrossRef]
- Ye, D.; Lin, Y.; Li, P.; Sun, M. Packed Levitated Marker for Entity and Relation Extraction. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics, (Volume 1: Long Papers), Dublin, Ireland, 22–27 May 2022; pp. 4904–4917. [Google Scholar]
- Yan, Z.; Yang, S.; Liu, W.; Tu, K. Joint Entity and Relation Extraction with Span Pruning and Hypergraph Neural Networks. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, Singapore, 6–10 December 2023; pp. 7512–7526. [Google Scholar]
- Li, Z.; Fu, L.; Wang, X.; Zhang, H.; Zhou, C. RFBFN: A relation-first blank filling network for joint relational triple extraction. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics: Student Research Workshop, Dublin, Ireland, 22–27 May 2022; pp. 10–20. [Google Scholar]
- Zheng, H.; Wen, R.; Chen, X.; Yang, Y.; Zhang, Y.; Zhang, Z.; Zhang, N.; Qin, B.; Ming, X.; Zheng, Y. PRGC: Potential Relation and Global Correspondence Based Joint Relational Triple Extraction. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), Online, 1–6 August 2021; pp. 6225–6235. [Google Scholar]
- Lu, Y.; Liu, Q.; Dai, D.; Xiao, X.; Lin, H.; Han, X.; Sun, L.; Wu, H. Unified Structure Generation for Universal Information Extraction. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Dublin, Ireland, 22–27 May 2022; pp. 5755–5772. [Google Scholar]
- Ngo, N.; Min, B.; Nguyen, T. Unsupervised domain adaptation for joint information extraction. In Findings of the Association for Computational Linguistics: EMNLP 2022; Association for Computational Linguistics: Abu Dhabi, United Arab Emirates, 2022; pp. 5894–5905. [Google Scholar]
- Van Nguyen, M.; Min, B.; Dernoncourt, F.; Nguyen, T. Learning cross-task dependencies for joint extraction of entities, events, event arguments, and relations. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, Abu Dhabi, United Arab Emirates, 7–11 December 2022; pp. 9349–9360. [Google Scholar]
- Li, Y.; Xue, B.; Zhang, R.; Zou, L. AtTGen: Attribute tree generation for real-world attribute joint extraction. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Toronto, ON, Canada, 9–14 July 2023; pp. 2139–2152. [Google Scholar]
- Zhang, R.; Li, Y.; Zou, L. A novel table-to-graph generation approach for document-level joint entity and relation extraction. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Toronto, ON, Canada, 9–14 July 2023; pp. 10853–10865. [Google Scholar]
- Kang, S.; Ji, L.; Liu, S.; Ding, Y. Cross-Lingual entity alignment model based on the similarities of entity descriptions and knowledge embeddings. Acta Electron. Sin. 2019, 47, 1841–1847. [Google Scholar]
- Zhou, L. Research on Intelligent Diagnosis and Maintenance Decision Method of Train Control On-Board Equipment in High-speed Railway. Ph.D. Thesis, Lanzhou Jiaotong University, Lanzhou, China, 2024. [Google Scholar]
- Jiang, T.; Bu, C.; Zhu, Y.; Wu, X. Two-Stage Entity Alignment: Combining Hybrid Knowledge Graph Embedding with Similarity-Based Relation Alignment. In Proceedings of the Pacific Rim International Conference on Artificial Intelligence, Yanuca Island, Fiji, 26–30 August 2019; pp. 162–175. [Google Scholar]
- Dettmers, T.; Minervini, P.; Stenetorp, P.; Riedel, S. Convolutional 2d knowledge graph embeddings. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018. [Google Scholar]
- Zhenbing, Z.; Jikun, D.; Yinghui, K.; Dongxia, Z. Construction and application of bolt and nut pair knowledge graph based on GGNN. Power Syst. Technol. 2020, 45, 98–106. [Google Scholar]
- Ren, H.; Yang, M.; Jiang, P. Improving attention network to realize joint extraction for the construction of equipment knowledge graph. Eng. Appl. Artif. Intell. 2023, 125, 106723. [Google Scholar] [CrossRef]
- Liu, P.; Qian, L.; Zhao, X.; Tao, B. The construction of knowledge graphs in the aviation assembly domain based on a joint knowledge extraction model. IEEE Access 2023, 11, 26483–26495. [Google Scholar] [CrossRef]
- Wang, T.; Zheng, L.; Lv, H.; Zhou, C.; Shen, Y.; Qiu, Q.; Li, Y.; Li, P.; Wang, G. A distributed joint extraction framework for sedimentological entities and relations with federated learning. Expert Syst. Appl. 2023, 213, 119216. [Google Scholar] [CrossRef]
- Devlin, J.; Chang, M.-W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), Minneapolis, MN, USA, 2–7 June 2019; Association for Computational Linguistics: Minneapolis, MN, USA, 2019; pp. 4171–4186. [Google Scholar]
- Li, L.; Ma, R.; Guo, Q.; Xue, X.; Qiu, X. BERT-ATTACK: Adversarial Attack Against BERT Using BERT. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), Online, 16–20 November 2020; pp. 6193–6202. [Google Scholar]
- Wang, Y.; Yu, B.; Zhang, Y.; Liu, T.; Zhu, H.; Sun, L. TPLinker: Single-stage Joint Extraction of Entities and Relations Through Token Pair Linking. In Proceedings of the 28th International Conference on Computational Linguistics, Barcelona, Spain, 8–13 December 2020; pp. 1572–1582. [Google Scholar]
- Theodoropoulos, C.; Moens, M.-F. An information extraction study: Take in mind the tokenization! In Proceedings of the Conference of the European Society for Fuzzy Logic and Technology, Palma, Spain, 4–8 September 2023; Springer Nature: Cham, Switzerland, 2023; pp. 593–606. [Google Scholar]
- Ma, Y.; Hiraoka, T.; Okazaki, N. Joint entity and relation extraction based on table labeling using convolutional neural networks. In Proceedings of the Sixth Workshop on Structured Prediction for NLP, Dublin, Ireland, 27 May 2022; Association for Computational Linguistics: Dublin, Ireland, 2022; pp. 11–21. [Google Scholar]
- Beltagy, I.; Lo, K.; Cohan, A. SciBERT: A Pretrained Language Model for Scientific Text. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China, 3–7 November 2019; pp. 3615–3620. [Google Scholar]
- Wang, J.; Lu, W. Two are Better than One: Joint Entity and Relation Extraction with Table-Sequence Encoders. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), Online, 16–20 November 2020; pp. 1706–1721. [Google Scholar]
- Eberts, M.; Ulges, A. Span-based joint entity and relation extraction with transformer pre-training. In ECAI 2020; IOS Press: Amsterdam, The Netherlands, 2020; pp. 2006–2013. [Google Scholar]
- Wei, Z.; Su, J.; Wang, Y.; Tian, Y.; Chang, Y. A Novel Cascade Binary Tagging Framework for Relational Triple Extraction. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020; pp. 1476–1488. [Google Scholar]
- Wang, Y.; Sun, C.; Wu, Y.; Yan, J.; Gao, P.; Xie, G. Pre-training entity relation encoder with intra-span and inter-span information. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), Online, 16–20 November 2020; pp. 1692–1705. [Google Scholar]
- Zhong, Z.; Chen, D. A Frustratingly Easy Approach for Entity and Relation Extraction. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Online, 6–11 June 2021; pp. 50–61. [Google Scholar]
- Sy, C.Y.; Maceda, L.L.; Canon, M.J.P.; Flores, N.M. Beyond BERT: Exploring the Efficacy of RoBERTa and ALBERT in Supervised Multiclass Text Classification. Int. J. Adv. Comput. Sci. Appl. 2024, 15, 223. [Google Scholar] [CrossRef]
- Sun, Y.; Wang, S.; Li, Y.; Feng, S.; Tian, H.; Wu, H.; Wang, H. Ernie 2.0: A continual pre-training framework for language understanding. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 8968–8975. [Google Scholar]
Entity 1 | Entity 2 | Characteristic | Alignment |
---|---|---|---|
locate in the fracture | on the fracture | Refer to the same position | True |
Pulling speed | Speed of pulling slab | Abbreviation | True |
increase equiaxed grain ratio | enhance equiaxed grain ratio | Refer to the same meaning | True |
enriched solute elements in the molten steel are drawn into voids | the voids draw in molten steel enriched with solute elements | Different word order | True |
serrated | curved | Different meaning | False |
S ≤ 0.006% | P ≤ 0.0125% | Different chemistry symbol | False |
Longitudinal cracks in slab | Longitudinal cracks | Have extra adjective phrase | False |
Longitudinal cracks on the surface | transverse cracks on the surface | Small character variations | False |
Parameter | Value | Parameter | Value |
---|---|---|---|
Entities | 1203 | Relations | 1193 |
Event Feature | 212 | Describe | 94 |
Casting Event | 425 | Lead to | 445 |
Quality Defect Solution | 154 | Feature is | 231 |
Defect Characteristics | 81 | Take action | 164 |
Quality Defect Name | 103 | Cause | 259 |
Cause of Quality Defect | 223 | Total | 2396 |
Parameter | SAPRM | Alignment | Explanation |
---|---|---|---|
epoch | 200 | 20 | number of train epoch |
batch size | 20 | 8 | number of samples in one batch |
lr | 2 × 10−5 | 2 × 10−5 | learning rate |
weight decay | 0 | 1 × 10−4 | weight decaying rate |
dropout | 0.1 | 0.1 | dropout rate for input |
clip | 0.25 | 0 | grad norm clip |
λ | 1 | - | proportion of NER and RE loss |
Models † | NER | RE | ||||
---|---|---|---|---|---|---|
P | R | F | P | R | F | |
TPLinker | - | - | - | 58.65 | 48.14 | 52.88 |
PRGC | - | - | - | 65.43 | 54.12 | 59.24 |
PFN | 84.98 | 79.95 | 82.39 | 76.79 | 64.84 | 70.31 |
TablERT-CNN | 85.64 | 84.79 | 85.10 | 67.20 | 70.07 | 68.54 |
SAPRM (Our Model) | 86.70 | 84.04 | 85.35 | 79.48 | 70.32 | 74.62 |
Ablation | Settings | NER | RE | ||||
---|---|---|---|---|---|---|---|
P | R | F | P | R | F | ||
Layers | N = 1 | 86.70 | 84.04 | 85.35 | 79.48 | 70.32 | 74.62 |
attention | w/o | 80.86 | 79.32 | 80.17 | 76.30 | 60.69 | 67.61 |
partition | w/o | 79.39 | 82.04 | 80.71 | 74.74 | 63.22 | 68.44 |
global | w/o | 82.85 | 79.76 | 81.27 | 75.62 | 61.65 | 67.94 |
λ | NER | RE | ||
---|---|---|---|---|
P | F | P | F | |
1 | 86.70 | 85.35 | 79.48 | 74.62 |
0.67 | 83.86 | 84.20 | 78.40 | 73.92 |
0.5 | 82.84 | 83.58 | 76.36 | 71.09 |
1.5 | 78.84 | 81.59 | 76.68 | 72.85 |
2 | 79.84 | 81.26 | 77.08 | 72.45 |
Models | ADE ◊⸸ | WebNLG ♦⸸ | SciERC ♦‡ | |||
---|---|---|---|---|---|---|
NER | RE | NER | RE | NER | RE | |
Table-Sequence [31] | 89.7 | 80.1 | - | - | - | - |
SpERT [32] | 89.3 | 79.2 | - | - | - | - |
PFN [28] | 89.6 | 80.0 | 98.0 | 93.6 | 66.8 | 38.4 |
Casrel [33] | - | - | 95.5 | 91.8 | - | - |
TPLinker [27] | - | - | - | 91.9 | - | - |
SPE [34] | - | - | - | - | 68.0 | 34.6 |
PURE [35] | - | - | - | - | 66.6 | 35.6 |
SAPRM (Our Model) | 90.3 | 82.7 | 98.1 | 92.7 | 70.2 | 39.3 |
Model | P | R | F |
---|---|---|---|
BERT [25] | 99.29 | 99.21 | 99.25 |
ALBERT [36] | 97.87 | 97.59 | 97.73 |
ERNIE [37] | 94.81 | 94.85 | 94.83 |
Structure | Precision | Recall | F1-Score |
---|---|---|---|
Last Layer | 98.94 | 98.82 | 98.88 |
Last 4 Layers | 97.87 | 97.68 | 97.78 |
Second to Last | 98.58 | 98.41 | 98.50 |
Last + First Layers | 98.23 | 98.06 | 98.14 |
Last 2 Layers | 98.23 | 98.00 | 98.11 |
Last 2 + First 2 Layers | 98.94 | 98.80 | 98.87 |
Weight Decay | Precision | Recall | F1-Score |
---|---|---|---|
0 | 98.94 | 98.82 | 98.88 |
5 × 10−2 | 97.87 | 87.58 | 97.73 |
5 × 10−3 | 98.23 | 97.97 | 98.10 |
5 × 10−4 | 98.58 | 98.39 | 98.48 |
1 × 10−4 | 99.29 | 99.21 | 99.25 |
5 × 10−5 | 98.23 | 97.97 | 98.10 |
Ablation | Settings | P | R | F |
---|---|---|---|---|
SSI and Attack | w/ | 99.29 | 99.21 | 99.25 |
Attack | w/o | 98.58 | 98.44 | 98.51 |
SSI and Attack | w/o | 97.90 | 87.32 | 92.22 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Wu, X.; She, Y.; Wang, X.; Lu, H.; Gao, Q. Domain-Specific Knowledge Graph for Quality Engineering of Continuous Casting: Joint Extraction-Based Construction and Adversarial Training Enhanced Alignment. Appl. Sci. 2025, 15, 5674. https://doi.org/10.3390/app15105674
Wu X, She Y, Wang X, Lu H, Gao Q. Domain-Specific Knowledge Graph for Quality Engineering of Continuous Casting: Joint Extraction-Based Construction and Adversarial Training Enhanced Alignment. Applied Sciences. 2025; 15(10):5674. https://doi.org/10.3390/app15105674
Chicago/Turabian StyleWu, Xiaojun, Yue She, Xinyi Wang, Hao Lu, and Qi Gao. 2025. "Domain-Specific Knowledge Graph for Quality Engineering of Continuous Casting: Joint Extraction-Based Construction and Adversarial Training Enhanced Alignment" Applied Sciences 15, no. 10: 5674. https://doi.org/10.3390/app15105674
APA StyleWu, X., She, Y., Wang, X., Lu, H., & Gao, Q. (2025). Domain-Specific Knowledge Graph for Quality Engineering of Continuous Casting: Joint Extraction-Based Construction and Adversarial Training Enhanced Alignment. Applied Sciences, 15(10), 5674. https://doi.org/10.3390/app15105674