You are currently viewing a new version of our website. To view the old version click .
Applied Sciences
  • Article
  • Open Access

20 June 2025

Root Cause Analysis of Cast Product Defects with Two-Branch Reasoning Network Based on Continuous Casting Quality Knowledge Graph

,
,
,
and
1
Shaanxi Joint Laboratory of Artificial Intelligence, School of Software Engineering, Xi’an Jiaotong University, Xi’an 710049, China
2
China National Heavy Machinery Research Institute Co., Ltd., Xi’an 710016, China
*
Authors to whom correspondence should be addressed.

Abstract

A variety of cast product defects may occur in the continuous casting process. By establishing a Continuous Casting Quality Knowledge Graph (C2Q-KG) focusing on the causes of cast product defects, enterprises can systematically sort out and express the relations between various production factors and cast product defects, which makes the reasoning process for the causes of cast product defects more objective and comprehensive. However, reasoning schemes for general KGs often use the same processing method to deal with different types of relations, without considering the difference in the number distribution of the head and tail entities in the relation, leading to a decrease in reasoning accuracy. In order to improve the reasoning accuracy of C2Q-KGs, this paper proposes a model based on a two-branch reasoning network. Our model classifies the continuous casting triples according to the number distribution of the head and tail entities in the relation and connects a two-branch reasoning network consisting of one connection layer and one capsule layer behind the convolutional layer. The connection layer is used to deal with the sparsely distributed entity-side reasoning task in the triple, while the capsule layer is used to deal with the densely distributed entity-side reasoning task in the triple. In addition, the Graph Attention Network (GAT) is introduced to enable our model to better capture the complex information hidden in the neighborhood of each entity and improve the overall reasoning accuracy. The experimental results show that compared with other cutting-edge methods on the continuous casting data set, our model significantly improves performance and infers more accurate root causes of cast product defects, which provides powerful guidance for enterprise production.

1. Introduction

In modern industrial manufacturing, continuous casting has been widely used in steel production. However, the continuous casting process is complex and involves many parameters and operation steps. Various cast product defects, such as cracks, segregation and pores, may occur, affecting quality of cast products and potentially causing production stagnation or equipment damage. Therefore, accurately identifying and analyzing the root causes of cast product defects are crucial.
Traditional cast product defect analysis mainly relies on experienced engineers who judge the causes of defects by observing the appearance of defects and combining their own experience [1]. This method is effective to a certain extent, but due to the complexity of the continuous casting process and the diversity of influencing factors, it is difficult to grasp all the potential defect causes comprehensively and systematically by manual experience alone. With the development of artificial intelligence and data analysis technology, how to use data-driven methods to analyze the root causes has become a research hotspot.
A knowledge graph (KG) is a model that represents knowledge in the form of a graph structure. It connects entities and relations in knowledge through nodes and edges, forming a networked knowledge system. A KG can not only express static relations among entities but also semantically reason about entities and their relations, so as to reveal hidden knowledge and information. With the development of technology, KGs have been widely used in various domains to form the so-called Domain-Specific KGs (DSKGs).
DSKGs can significantly improve the effect of information retrieval, knowledge discovery, decision support and other tasks in a specific domain [2]. For example, in the medical domain, DSKGs can help doctors diagnose diseases [3] and perform medical research and clinical practice [4] more accurately by integrating patient data. In the manufacturing domain, DSKGs can optimize production processes, predict equipment failures and improve product quality [5]. In the financial domain, DSKGs can assist in risk assessment and intelligent investment advisory [6].
In the root cause analysis of cast product defects, the Continuous Casting Quality Knowledge Graph (C2Q-KG) can integrate different process parameters, equipment status and environmental factors in the continuous casting process. Through the systematic analysis of the correlation of this information, the causes of cast product defects can be reasoned out more objectively.
However, existing knowledge graph reasoning methods often use the same reasoning method when dealing with different types of relation features and fail to fully consider the difference in the number distribution of head and tail entities in the relation. Applying the same treatment leads to uneven reasoning accuracy, and the reasoning performance is often unsatisfactory.
In order to solve this problem, this paper proposes a knowledge graph reasoning model based on a two-branch reasoning network (KGTBRN), which aims to improve the accuracy of root cause analysis for cast product defects and provide powerful decision support for quality control and process optimization in production for continuous casting enterprises.
Our contributions in this paper are as follows:
  • We propose KGTBRN in the continuous casting domain. Aiming at the requirements of root cause analysis for cast product defects, one connection layer and one capsule layer are used to deal with different types of entity-side reasoning tasks, so as to improve both the adaptability of KGTBRN to different relation features and the reasoning accuracy.
  • KGTBRN is experimentally verified on a continuous casting data set and compared with existing cutting-edge methods. The experimental results show that KGTBRN obtains the best mean reciprocal rank and the highest HITS@1, HITS@3 and HITS@10 on the continuous casting data set. It also demonstrates clear advantages on the benchmark data set FB15k-237 and shows good generalization performance on the WN18RR data set.
The rest of the paper is structured as follows: Section 2 reviews the related works on knowledge graph reasoning and Graph Attention Network (GAT). Section 3 introduces the proposed knowledge graph reasoning model, KGTBRN. Section 4 introduces the C2Q-KG and analyzes the experimental results. Section 5 gives the conclusions of this paper.

3. Methods

3.1. Problem Modeling

The task of the root cause analysis of cast product defects is formalized as a link prediction task of head entity or tail entity. The C2Q-KG is expressed in the form of triples as G = h , r , t , where h , t E represent the continuous casting head entity and the continuous casting tail entity, respectively, and r R represents the continuous casting relation between the continuous casting head entity and the tail entity. For a cast product defect head entity h and a continuous casting relation r, the triple is represented as d e f e c t   h , r , t , and its reasoning task is to reason on the root cause tail entity t. On the other hand, for a continuous casting relation r and a cast product defect tail entity t, the triple is represented as h , r , d e f e c t   t , and its reasoning task is to reason on the root cause head entity h.

3.2. Continuous Casting Triple-Embedding Module Based on GAT

Aiming at the existing C2Q-KG, we propose a continuous casting triple-embedding module based on the GAT, which makes KGTBRN better capture the complex information hidden in the neighborhood of each entity [21] and improves the accuracy of the subsequent reasoning layer. Specifically, this module consists of two parts: the continuous casting relation classification module and the GAT embedding module.

3.2.1. Continuous Casting Relation Classification Module

This module realizes the preprocessing task of continuous casting triples in the C2Q-KG and counts the number of entities and the number of different relations contained in it. Then, it classifies the relations according to the number distribution of head and tail entities. Specifically, the processing method for the classification of relations is detailed below.
The average number of continuous casting tail entities η t for each continuous casting head entity and the average number of continuous casting head entities η h for each continuous casting tail entity are counted. For a continuous casting triple h , r , t , if η h < 1.5 and η t < 1.5 , the relation r is classified as a 1-1 type; if η h < 1.5 and η t 1.5 , the relation r is classified as a 1-N type; if η h 1.5 and η t 1.5 , the relation r is classified as an N-1 type; and if η h 1.5 and η t 1.5 , the relation r is classified as an N-N type.
In the subsequent reasoning task of triples, different types of relations are classified to improve the accuracy of reasoning about the causes of cast product defects on the C2Q-KG, and the problem that the existing knowledge graph reasoning methods cannot balance the reasoning accuracy of all continuous casting relations is solved.

3.2.2. GAT Embedding Module

In this module, we input the relations obtained by the continuous casting classification module and the feature vectors of its neighborhood into the GAT. Then, we capture the hidden neighborhood information between the entities and relations through the two-layer GAT. Finally, we obtain the updated embedding vectors of head entity, tail entity and relation. Figure 1 shows the structural diagram of this module.
Figure 1. The embedding module structure of the two-layer GAT.
For a continuous casting triple τ = h , r , t , similar to Nathani et al. [35], we first obtain the representation of each continuous casting triple associated with the continuous casting head entity h, tail entity t and relation r by performing a linear transformation on the cascade of its three feature vectors, as shown in Equation (1).
M h r t = W 1 v h v r v t ,
where v h , v r ,   v t denote the embeddings of head entity h , tail entity t and relation   r , respectively; M h r t   is the embedding vector representation of a continuous casting triple τ ; and W 1 denotes the linear transformation matrix.
We perform a linear transformation on M h r t , and the attention coefficient b h r t of a specific triple τ is obtained through the nonlinear activation function LeakyReLU. b h r t represents the importance of the continuous casting triple in the cast product defect reasoning task, as shown in Equation (2). LeakyReLU is determined based on common practice in the existing literature and empirical performance during preliminary experiments. Then, we perform the normalization operation on b h r t and get the normalized attention coefficient a h r t , as shown in Equation (3).
b h r t = L e a k y R e L U ( W 2 M h r t ) ,
a h r t = s o f t m a x ( b h r t ) = e x p ( b h r t ) e E t r R t e x p ( b h r t ) ,
where E τ denotes the set of all neighborhood of the head entity h and R τ   represents the set of relations r connecting the entity in E τ and head entity h .
After obtaining the normalized attention coefficient a h r t , the new head entity embedding vector h is calculated by weighting, as shown in Equation (4). The attention values of each head are averaged and updated to obtain the final head entity embedding vector h , as shown in Equation (5).
h = σ ( t E τ r R τ a h r t M h r t ) .
h = σ ( 1 N n = 1 N t E τ r R τ a h r t n M h r t n ) .
Similarly, the same update method is used to obtain the final tail embedding vector t .
This module uses N independent attention mechanisms to calculate the embedding vectors of continuous casting triples and concatenates these embedding results to obtain a more expressive continuous casting triple-embedding representation.

3.3. Cast Product Defect Root Cause Reasoning Module of C2Q-KG Based on Two-Branch Reasoning Network

This module reasons on the root causes of cast product defects on the basis of the updated continuous casting triple-embedding vector obtained previously. It is composed of a convolutional layer and a reasoning layer and uses a flexible reasoning network to reason on the root causes for the different types of relations previously divided. We modify the convolutional layers of the ConvKB model [36]. The reasoning layer includes a two-branch network including a connection layer and a capsule layer.
The convolutional layer connects the updated embedding vectors of head entity, tail entity and relation to obtain the embedding matrix. Then, the Convolutional Neural Network (CNN) is used to perform a convolution operation. The structure is shown in Figure 2.
Figure 2. The structure of the convolutional layer.
The embedding form ( v h , v r , v t ) of each continuous casting triple ( h , r , t ) is regarded as a continuous casting matrix A = V h , V r , V t R k × 3 , where A i   denotes the row of A and k × 3 denotes the dimension size of ( h , r , t ) . Referring to the transfer characteristics based on the transfer model, the convolutional layer uses filters of ω   R 1 × 3 size to study the overall relation between the embedded continuous casting triples ( v h , v r , v t ) in the same dimension. The filter ω repeats the operation on each row of A and finally generates the set of feature maps v = v 1 , v 2 , , v k R k , as shown in Equation (6).
v i = R e L U ( w A i + b ) ,
where b R is a bias term.
For the KG link prediction task, CapSe [37] uses the capsule network to optimize accuracy. However, the capsule network only shows good performance when reasoning on the multi-terminal entity (N-side) of the triple, and its effect on the single-terminal entity (1-side) is not as good as those connection models. Therefore, we use the capsule layer, which consists of two capsule networks, to deal with the N-side continuous casting entity reasoning tasks, and use the connection layer to deal with the 1-side continuous casting entity reasoning tasks.
Through the two-branch reasoning network, the overall reasoning process of the root causes is refined, so that KGTBRN can better adapt to the diversified characteristics of relations in the continuous casting domain and improve the reasoning accuracy. Figure 3 shows the structure diagram of the two-branch reasoning network.
Figure 3. The structure of the two-branch reasoning network.
For 1-side continuous casting entity reasoning tasks (including the 1-1 and 1-N types of head entity reasoning tasks and the 1-1 and N-1 types of tail entity reasoning tasks), the feature-mapping vector can be input into the connection layer directly to form the output vector e R τ k × 1 , where τ is the number of filters and k is the embedding dimension of the vector in the convolutional layer.
For N-side continuous casting entity reasoning tasks (including the N-N and N-1 types of head entity reasoning tasks and the N-N and 1-N types of tail entity reasoning tasks), the feature-mapping vector can be reconstructed into the corresponding first capsule layer. It is reconstructed as a matrix and routed to the second capsule layer to generate the continuous casting triple vector compressed by the squash function.
K capsules are constructed in the first capsule layer, where the entries of the same dimension from all the continuous casting triple feature-mapping vectors are encapsulated into the same capsule, so that each capsule can capture many features between the entries of the corresponding dimension in the embedded continuous casting triple. These features are re-summarized through the routing process and squash function compression and finally recorded in the second capsule network.
Additionally, the routing process between the two capsule layers is as follows: Each capsule vector u i of the first capsule layer is multiplied by a weight matrix W i to obtain the updated capsule vector u i ^ , as shown in Equation (7).
u i ^ = W i u i .
The weighted sum of all capsule vectors is performed, where the weighting coefficient c i is obtained by the output of softmax function in the first capsule layer. The routing result s is obtained by weighted summation, and the final continuous casting triple output vector of this layer is obtained by the squash function, as shown in Equation (8).
s = i c i u i ^ , e = s q u a s h ( s ) ,
where
s q u a s h ( s ) = s 2 1 + s 2 s s .
The modular length of the output vector e obtained by the above two reasoning methods is the reasoning accuracy of the continuous casting triple. The scoring function f is operated by computing the modulus of e. Concat(*) denotes the connection layer, and capsnet(*) denotes the capsule layer, as shown in Equation (9).
f ( h , r , t ) = e ,
where
e = c o n c a t ( g ( v h , v r , v t     ω ) ) c a p s n e t ( g ( v h , v r , v t     ω ) ) .
KGTBRN is trained by minimizing the loss function and is optimized by using the Adam optimizer. The loss function is as shown Equation (10), where G is the set of invalid continuous casting triples, which are generated by the head entity or tail entity that damages the triples in the valid set G .
L = ( h , r , t ) { G G } log ( 1 + exp ( t ( h , r , t ) f ( h , r , t ) ) ) ,
where
t ( h , r , t ) = 1 , w h e n ( h , r , t ) G 1 , w h e n ( h , r , t ) G .
The overall structure of KGTBRN is shown in Figure 4. KGTBRN consists of two main components: the embedding module and the reasoning module. The embedding module includes a continuous casting relation classification module and a two-layer GAT, which together capture structural and semantic information from triples in the C2Q-KG. The reasoning module includes a convolutional layer and a two-branch reasoning network, designed to handle different distributed entity-side reasoning task and improve the accuracy of KGTBRN.
Figure 4. The overall structure of KGTBRN. The numbers indicate the steps of KGTBRN.
The specific classification of relations and the reasoning selection examples of the two-branch reasoning network are shown in Table 1.
Table 1. Relation classification and selection examples of two-branch reasoning network.
According to the different types of relation obtained by the previous classification and the needs of the reasoning tasks, we randomly select another entity that is not equal to the current entity from the entity set for replacement to generate negative samples of continuous casting triples for model training. For example, for the 1-N type continuous casting triple, the following steps are necessary:
(i)
Replace the existing head entity, and retain the relation and tail entity to generate a negative sample of the 1-N type, which is used for training the reasoning tasks of the head entity in the 1-N type casting triple.
(ii)
Replace the existing tail entity, and retain the relation and head entity to generate a negative sample of the 1-N type, which is used for training the reasoning tasks of the tail entity in the 1-N type casting triple.
The above two steps are performed for all continuous casting triples to generate a set of all negative samples as the invalid continuous casting triple set G .

4. Experiment and Analysis

4.1. C2Q-KG

The detailed metadata of the C2Q-KG used in this paper is shown in Table 2. The C2Q-KG contains 1626 continuous casting triples, involving 1144 entities and 5 types of relations.
Table 2. The detailed metadata of the C2Q-KG.
The “describe” relation denotes basic information about cast product defects, such as their location. The “feature is” relation denotes specific manifestations and related attributes of cast product defects, such as depth and temperature. The “cause” relation denotes that a certain root cause leads to a dynamic cast product defect phenomenon, while the “lead to” relation denotes a cause resulting in a static cast product defect. The “solution” relation denotes the optimization measures adopted for cast product defects.
These relations cover the four types of 1-1, 1-N, N-1 and N-N. The reasoning accuracy and stability of KGTBRN can be measured by using the standard knowledge graph reasoning evaluation indicators in the domain. The main indicators include hit probability among the top k results (HITS@K) and mean reciprocal rank (MRR), as shown in Equations (11) and (12), where rank i denotes the rank of the correct entity among all reasoning results and S denotes the number of data in the set.
H I T S @ K = 1 S i = 1 s I ( r a n k i k )
M R R = 1 S i = 1 s 1 r a n k i

4.2. Performance Evaluation of Root Cause Analysis Model for Cast Product Defects

Table 3 shows the reasoning accuracy of KGTBRN on the continuous casting data set and the comparison with TransE [38], DistMult [39], ComplEX [40], KGLM [41] and KERMIT [42].
Table 3. Comparison of reasoning accuracy of each model on continuous casting data set.
On the task of the root cause analysis of cast product defects in the C2Q-KG, KGTBRN shows the highest accuracy. On the continuous casting data set, KGTBRN has the best performance on MRR, HITS@1, HITS@3 and HITS@10.
Compared with the comparison models, the improvement in HITS@1 means that the differentiated reasoning mode for different relation types adopted by KGTBRN is more in line with the diversified characteristics of triple relation types in the C2Q-KG. KGTBRN scores a 6.9% higher value than the optimal model KERMIT in HITS@1, and 6.2% and 5.6% higher values in HITS@3 and HITS@10, respectively. These findings prove that KGTBRN is effective in the task of the root cause analysis of cast product defects.

4.3. Ablation Experiment of Root Cause Analysis Model for Cast Product Defects

Table 4 shows the results of the ablation experiment of KGTBRN on the task of the root cause analysis of cast product defects. “Model-attention” is the model where the GAT layer of KGTBRN is removed, and only the two-branch reasoning network is used for the root cause analysis task. “Model-concat” is the model where the capsule layer in the two-branch reasoning network of KGTBRN is removed, and “Model-capsnet” is the model where the connection layer in the two-branch reasoning network of KGTBRN is removed.
Table 4. Ablation experiment results of KGTBRN on the task of the root cause analysis of cast product defects.
It can be seen from Table 4 that the GAT layer, the connection layer and the capsule layer improve the reasoning accuracy. Taking the HITS@1 index as an example, the addition of the GAT layer improves HITS@1 by 5.7%. Using the connection layer to process the 1-side reasoning task improves HITS@1 by 2%, and using the two capsule layers to process the N-side reasoning task improves HITS@1 by 4.6%.
This shows that the two-branch reasoning network uses the idea of “classifying relation types, and using different reasoning methods to deal with different types of relations separately” is more suitable for the C2Q-KG. Meanwhile, the GAT layer can better identify the potential feature relations between relations and entities, so as to improve the reasoning accuracy of the overall model.

4.4. Experiments on Hyperparameters of GAT Layer and Reasoning Layer

Table 5 shows the experiment results of the hyperparameters of the GAT layer and the reasoning layer. In Table 5, “epoch_attention” represents the iteration number of the GAT layer and takes values of {500, 800, 1000}, and “epoch_reasoning” represents the iteration number of the reasoning layer and takes values of {150, 200, 250}. Experiments are carried out on these two hyperparameter permutations and combinations. When “epoch_attention” is 1000 and “epoch_reasoning” is 200, the reasoning accuracy of KGTBRN reaches the highest.
Table 5. Experimental results of hyperparameters of GAT layer and reasoning layer.
Figure 5 shows the loss iteration values of the GAT layer and the two-branch reasoning network of KGTBRN with the optimal parameters, respectively. It can be seen that all parts have converged successfully, proving that both the GAT layer and the two-branch reasoning network successfully capture the knowledge information of the C2Q-KG.
Figure 5. Loss iteration values of the GAT layer (a) and the two-branch reasoning network (b).

4.5. Generalization Experiment of Root Cause Analysis Model for Cast Product Defects

Table 6 and Table 7 show the experimental results of KGTBRN on the benchmark data sets FB15K-237 and WN18RR, respectively. The selected comparison models are ConvE [43], TransE [38], ConvKB [36], KGLM [41], KERMIT [42] and NBFNET [44].
Table 6. Experimental results on data set FB15K-237.
Table 7. Experimental results on data set WN18RR.
It can be seen that on the FB15K-237 data set, KGTBRN achieves the highest HITS@1 and HITS@3 and is suboptimal in HITS@10. Also, KGTBRN achieves the highest HITS@3 on WN18RR, indicating that the two-branch reasoning network can improve the overall reasoning accuracy of the KG by effectively handling different entity distribution patterns. Table 6 and Table 7 further reflect the robustness and generalization capability of KGTBRN.
Take FB15K-237 as an example, analyzing the triple distribution of and continuous casting data set, the results are shown in Table 8. The statistical results show that the proportions of 1-1-type and 1-N-type triples in the FB15K-237 data set are small: the 1-1-type triples represent only 0.94% and the 1-N-type ones only 6.32%. However, the distribution on the continuous casting data set is more uniform, so it is speculated that the uneven distribution of triples in the FB15K-237 affects the reasoning accuracy of the 1-side of KGTBRN; then, the overall reasoning results are reduced in the HITS@10 index, achieving suboptimal results.
Table 8. Statistical results of triple distribution of FB15K-237 and continuous casting data set.
Table 9 shows examples of actual reasoning results. According to the needs of the reasoning task, KGTBRN can use different reasoning methods according to the 1-side and N-side and give the top two root causes obtained by reasoning.
Table 9. Examples of actual reasoning results of root cause analysis task.
When researchers need to reason about the root cause based on the appearance of a certain cast product defect, they can use “cause” as the input relation of the root cause analysis task. When researchers need to reason about the root cause based on a certain production phenomenon, “lead to” can be used as the input relation.
From the first row of Table 9, it can be seen that when cracks appear on the surface of the casting, KGTBRN gives the top two root causes as “the first brittleness temperature zone moves towards low temperature” and “deviation in critical stress appears”. Based on these results, workers can make the corresponding adjustments to relevant continuous casting process parameters.
Figure 6 illustrates the cause reasoning results for “Crack”. KGTBRN identifies eight potential root causes for it, including “Stress concentration”, ”Shell unevenness”, ”Critical stress drop”, etc.
Figure 6. Reasoning results for cause of “Crack” defect.

5. Conclusions

This paper proposes a reasoning model of a C2Q-KG based on a GAT and a two-branch reasoning network. Firstly, KGTBRN classifies the continuous casting triples according to the number distribution of the head entities and tail entities for each relation in the C2Q-KG. For the triples with a sparse entity distribution (1-side), KGTBRN uses the connection layer to ensure that the relevant entities can still be accurately reasoned on in the case of lacking information. For the triples with a dense entity distribution (N-side), the capsule layer is used for reasoning.
In order to further improve the performance of KGTBRN, KGTBRN also introduces a GAT. By dynamically adjusting the attention distribution in different neighborhoods, the GAT enhances the ability of KGTBRN to capture the local neighborhood information of triples and enables it to better understand the implicit connections among entities, so that KGTBRN can still maintain high reasoning accuracy in the face of complex relations.
In the experimental part, we compare KGTBRN with other cutting-edge models on the C2Q-KG. The experiment results show that KGTBRN shows higher accuracy and reliability, improving by 6.9% on HITS@1, by 6.2% on HITS@3 and by 5.6% on HITS@10. This shows that KGTBRN can not only adapt to reasoning tasks under different entity distributions but also perform effective root cause analysis of cast product defects in complex continuous casting process scenarios.
We provide an effective knowledge graph reasoning model for the continuous casting domain. By combining the GAT and the two-branch reasoning network, the influence of the unbalanced distribution of entities on the reasoning accuracy is successfully solved. As KGTBRN is constructed based on the data characteristics of the C2Q-KG in the continuous casting domain, generalizing it to other industrial domains still presents challenges. Additionally, the two-branch reasoning network may increase model complexity and implementation difficulty. Future work will explore broader applications and introduce quality evaluation metrics to enhance the generalization capability of KGTBRN.
The application of this method can help enterprises to more accurately identify the potential root causes of cast product defects in the production process, thereby advancing quality management technologies in intelligent manufacturing. With strong scalability and application potential, it is expected to promote process optimization and product quality improvement, thus facilitating the intelligent and digital transformation of the manufacturing industry.

Author Contributions

Conceptualization, X.W. (Xiaojun Wu); methodology, X.W. (Xinyi Wang); software, X.W. (Xinyi Wang), M.S. and Y.S.; validation, X.W. (Xiaojun Wu), X.W. (Xinyi Wang), Y.S. and M.S.; formal analysis, X.W. (Xiaojun Wu); investigation, Q.G.; resources, Q.G.; data curation, X.W. (Xiaojun Wu), X.W. (Xinyi Wang), Y.S. and Q.G.; writing—original draft preparation, X.W. (Xinyi Wang); writing—review and editing, X.W. (Xiaojun Wu), Y.S. and M.S.; visualization, X.W. (Xiaojun Wu); supervision, X.W. (Xiaojun Wu); project administration, X.W. (Xiaojun Wu); funding acquisition, Q.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research study was funded by Shaanxi Province Innovation Capacity Support Plan 2024RS-CXTD-23.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data presented in this study are available upon request from the corresponding author.

Acknowledgments

The authors would like to thank the anonymous reviewers for their valuable feedback.

Conflicts of Interest

Author Qi Gao was employed by the company China National Heavy Machinery Research Institute Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Xia, L.; Liang, Y.; Leng, J.; Zheng, P. Maintenance Planning Recommendation of Complex Industrial Equipment Based on Knowledge Graph and Graph Neural Network. Reliab. Eng. Syst. Saf. 2023, 232, 109068. [Google Scholar] [CrossRef]
  2. Birs, I.R.; Muresan, C.; Copot, D.; Ionescu, C.-M. Model identification and control of electromagnetic actuation in continuous casting process with improved quality. IEEE-CAA J. Autom. Sin. 2023, 10, 203–215. [Google Scholar] [CrossRef]
  3. Chandak, P.; Huang, K.; Zitnik, M. Building a knowledge graph to enable precision medicine. Sci. Data 2023, 10, 67. [Google Scholar] [CrossRef]
  4. Liu, Z.; Xiao, L.; Chen, J.; Song, L.; Qi, P.; Tong, Z. A Multimodal Knowledge Graph for Medical Decision Making Centred Around Personal Values. In Proceedings of the 2023 26th International Conference on Computer Supported Cooperative Work in Design (CSCWD), Rio de Janeiro, Brazil, 24–26 May 2023; pp. 1638–1643. [Google Scholar]
  5. Abu-Salih, B. Domain-specific knowledge graphs: A survey. J. Netw. Comput. Appl. 2021, 185, 103076. [Google Scholar] [CrossRef]
  6. Zhang, Z.; Ni, Z.; Liu, Z.; Xia, S. Predicting Dynamic Relationship for Financial Knowledge Graph. Data Anal. Knowl. Discov. 2023, 7, 39–50. [Google Scholar] [CrossRef]
  7. Fani, V.; Antomarioni, S.; Bandinelli, R.; Ciarapica, F.E. Data Mining and Augmented Reality: An Application to the Fashion Industry. Appl. Sci. 2023, 13, 2317. [Google Scholar] [CrossRef]
  8. Liu, S.; Lyu, Q.; Liu, X.; Sun, Y. Synthetically Predicting the Quality Index of Sinter Using Machine Learning Model. Ironmak. Steelmak. 2020, 47, 828–836. [Google Scholar]
  9. Lee, S.Y.; Tama, B.A.; Choi, C.; Hwang, J.-Y.; Bang, J.; Lee, S. Spatial and Sequential Deep Learning Approach for Predicting Temperature Distribution in a Steel-Making Continuous Casting Process. IEEE Access 2020, 8, 21953–21965. [Google Scholar] [CrossRef]
  10. Zhang, J.; Ling, W. Quality Defect Tracing of Hot Rolled Strip Based on Knowledge Graph Reasoning. Comput. Integr. Manuf. Syst. 2024, 30, 1105. [Google Scholar]
  11. Wu, X.; Kang, H.; Yuan, S.; Jiang, W.; Gao, Q.; Mi, J. Anomaly Detection of Liquid Level in Mold During Continuous Casting by Using Forecasting and Error Generation. Appl. Sci. 2023, 13, 7457. [Google Scholar] [CrossRef]
  12. Takalo-Mattila, J.; Heiskanen, M.; Kyllonen, V.; Maatta, L.; Bogdanoff, A. Explainable Steel Quality Prediction System Based on Gradient Boosting Decision Trees. IEEE Access 2022, 10, 68099–68110. [Google Scholar] [CrossRef]
  13. Li, Y.; Li, Y.; Liu, J.; Fan, Z.; Wang, Q. Research on Segmentation of Steel Surface Defect Images Based on Improved Res-UNet Network. J. Electron. Inf. Technol. 2022, 44, 1513–1520. [Google Scholar]
  14. Cao, Y.; Jia, F.; Manogaran, G. Efficient Traceability Systems of Steel Products Using Blockchain-Based Industrial Internet of Things. IEEE Trans. Ind. Inform. 2020, 16, 6004–6012. [Google Scholar] [CrossRef]
  15. Kim, J.; Kwon, Y.; Jo, Y.; Choi, E. KG-GPT: A General Framework for Reasoning on Knowledge Graphs Using Large Language Models. In Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP), Singapore, 6–10 December 2023; pp. 9410–9421. [Google Scholar]
  16. Wu, T.; Qi, G.; Li, C.; Wang, M. A Survey of Techniques for Constructing Chinese Knowledge Graphs and Their Applications. Sustainability 2018, 10, 3245. [Google Scholar] [CrossRef]
  17. Wu, X.; She, Y.; Wang, X.; Lu, H.; Gao, Q. Domain-Specific Knowledge Graph for Quality Engineering of Continuous Casting: Joint Extraction-Based Construction and Adversarial Training Enhanced Alignment. Appl. Sci. 2025, 15, 5674. [Google Scholar] [CrossRef]
  18. Zhou, L.; Pan, M.; Sikorski, J.J.; Garud, S.; Aditya, L.K.; Kleinelanghorst, M.J.; Karimi, I.A.; Kraft, M. Towards an Ontological Infrastructure for Chemical Process Simulation and Optimization in the Context of Eco-Industrial Parks. Appl. Energy 2017, 204, 1284–1298. [Google Scholar] [CrossRef]
  19. Chen, Z.; Liu, Y.; Valera-Medina, A.; Robinson, F. Multi-Sourced Modelling for Strip Breakage Using Knowledge Graph Embeddings. Procedia CIRP 2021, 104, 1884–1889. [Google Scholar] [CrossRef]
  20. Zheng, P.; Xia, L.; Li, C.; Li, X.; Liu, B. Towards Self-X cognitive manufacturing network: An industrial knowledge graph-based multi-agent reinforcement learning approach. J. Manuf. Syst. 2021, 61, 16–26. [Google Scholar] [CrossRef]
  21. Mou, T.; Li, S. Knowledge Graph Construction for Control Systems in Process Industry. Chin. J. Intell. Sci. Technol 2022, 4, 129–141. [Google Scholar]
  22. Wu, Z.; Pan, S.; Chen, F.; Long, G.; Zhang, C.; Yu, P.S. A Comprehensive Survey on Graph Neural Networks. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 4–24. [Google Scholar] [CrossRef]
  23. Veličković, P.; Cucurull, G.; Casanova, A.; Romero, A.; Lio, P.; Bengio, Y. Graph Attention Networks. In Proceedings of the Sixth International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  24. Wang, Q.; Jiang, H.; Yi, S.; Yang, L.; Nai, H.; Nie, Q. Hyperbolic Representation Learning for Complex Networks. J. Softw. 2020, 32, 93–117. [Google Scholar]
  25. Wang, X.; Ji, H.; Shi, C.; Wang, B.; Ye, Y.; Cui, P.; Yu, P.S. Heterogeneous Graph Attention Network. In Proceedings of the World Wide Web Conference, San Francisco, CA, USA, 13–17 May 2019; pp. 2022–2032. [Google Scholar]
  26. Wang, Q.; Mao, Z.; Wang, B.; Guo, L. Knowledge Graph Embedding: A Survey of Approaches and Applications. IEEE Trans. Knowl. Data Eng. 2017, 29, 2724–2743. [Google Scholar] [CrossRef]
  27. Deng, W.; Zhang, Y.; Yu, H.; Li, H. Knowledge Graph Embedding Based on Dynamic Adaptive Atrous Convolution and Attention Mechanism for Link Prediction. Inf. Process. Manag. 2024, 61, 103642. [Google Scholar] [CrossRef]
  28. Sun, Z.; Deng, Z.-H.; Nie, J.-Y.; Tang, J. Rotate: Knowledge graph embedding by relational rotation in complex space. In Proceedings of the Seventh International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  29. Abboud, R.; Ceylan, I.; Lukasiewicz, T.; Salvatori, T. Boxe: A Box Embedding Model for Knowledge Base Completion. Adv. Neural Inf. Process. Syst. 2020, 33, 9649–9661. [Google Scholar]
  30. Niu, G.; Li, B.; Zhang, Y.; Sheng, Y.; Shi, C.; Li, J.; Pu, S. Joint Semantics and Data-Driven Path Representation for Knowledge Graph Reasoning. Neurocomputing 2022, 483, 249–261. [Google Scholar] [CrossRef]
  31. Le, T.; Le, N.; Le, B. Knowledge Graph Embedding by Relational Rotation and Complex Convolution for Link Prediction. Expert Syst. Appl. 2023, 214, 119–122. [Google Scholar] [CrossRef]
  32. Wang, Z.; Li, L.; Zeng, D. Srgcn: Graph-Based Multi-Hop Reasoning on Knowledge Graphs. Neurocomputing 2021, 454, 280–290. [Google Scholar] [CrossRef]
  33. Junhua, D.; Yucheng, H.; Yi-an, Z.; Dong, Z. Attention-Based Relational Graph Convolutional Network for Knowledge Graph Reasoning. In Proceedings of the 2022 21st International Symposium on Communications and Information Technologies (ISCIT), Xi’an, China, 27–30 September 2022; pp. 216–221. [Google Scholar]
  34. Marra, G.; Diligenti, M.; Giannini, F. Relational reasoning networks. Knowl.-Based Syst. 2025, 310, 112822. [Google Scholar] [CrossRef]
  35. Nathani, D.; Chauhan, J.; Sharma, C.; Kaul, M. Learning Attention-based Embeddings for Relation Prediction in Knowledge Graphs. In Proceedings of the 57th Annual Meeting of the Association-for-Computational-Linguistics (ACL), Florence, Italy, 28 July–2 August 2019; pp. 4710–4723. [Google Scholar]
  36. Nguyen, D.Q.; Nguyen, T.D.; Nguyen, D.Q.; Phung, D. A Novel Embedding Model for Knowledge Base Completion Based on Convolutional Neural Network. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, New Orleans, LA, USA, 1–6 June 2018; pp. 327–333. [Google Scholar]
  37. Nguyen, D.Q.; Vu, T.; Nguyen, T.D.; Nguyen, D.Q.; Phung, D.; Assoc Computat, L. A Capsule Network-based Embedding Model for Knowledge Graph Completion and Search Personalization. In Proceedings of the Conference of the North-American-Chapter of the Association-for-Computational-Linguistics—Human Language Technologies (NAACL-HLT), Minneapolis, MN, USA, 2–7 June 2019; pp. 2180–2189. [Google Scholar]
  38. Bordes, A.; Usunier, N.; Garcia-Duran, A.; Weston, J.; Yakhnenko, O. Translating Embeddings for Modeling Multi-Relational Data. Adv. Neural Inf. Process. Syst. 2013, 26, 26. [Google Scholar]
  39. Yang, B.; Yih, W.-t.; He, X.; Gao, J.; Deng, L. Embedding Entities and Relations for Learning and Inference in Knowledge Bases. In Proceedings of the International Conference on Learning Representations, Banff, AB, Canada, 14–16 April 2014; pp. 2181–2187. [Google Scholar]
  40. Trouillon, T.; Welbl, J.; Riedel, S.; Gaussier, É.; Bouchard, G. Complex Embeddings for Simple Link Prediction. In Proceedings of the International Conference on Machine Learning, New York, NY, USA, 19–24 June 2016; pp. 2071–2080. [Google Scholar]
  41. Youn, J.; Tagkopoulos, I. KGLM: Integrating Knowledge Graph Structure in Language Models for Link Prediction. In Proceedings of the STARSEM, Seattle, WA, USA, 14–15 July 2022. [Google Scholar]
  42. Li, H.; Yu, B.; Wei, Y.; Wang, K.; Da Xu, R.Y.; Wang, B. Kermit: Knowledge Graph Completion of Enhanced Relation Modeling with Inverse Transformation. arXiv 2023, arXiv:2309.14770. [Google Scholar] [CrossRef]
  43. Dettmers, T.; Minervini, P.; Stenetorp, P.; Riedel, S. Convolutional 2D Knowledge Graph Embeddings. In Proceedings of the 32nd AAAI Conference on Artificial Intelligence/30th Innovative Applications of Artificial Intelligence Conference/8th AAAI Symposium on Educational Advances in Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; pp. 1811–1818. [Google Scholar]
  44. Zhu, Z.; Zhang, Z.; Xhonneux, L.-P.; Tang, J. Neural Bellman-Ford Networks: A General Graph Neural Network Framework for Link Prediction. In Proceedings of the 35th Annual Conference on Neural Information Processing Systems (NeurIPS), Electr Network, Online, 6–14 December 2021; pp. 29476–29490. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.