Next Article in Journal
Ultrasonic Guided Wave Health Monitoring of High-Temperature Aircraft Structures Based on Variational Mode Decomposition and Fuzzy Entropy
Previous Article in Journal
Numerical and Experimental Study of a Wearable Exo-Glove for Telerehabilitation Application Using Shape Memory Alloy Actuators
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Knowledge Graph Completion for High-Speed Railway Turnout Switch Machine Maintenance Based on the Multi-Level KBGC Model

1
School of Automation and Electrical Engineering, Lanzhou Jiaotong University, Lanzhou 730070, China
2
CRRC Zhuzhou Institute Co., Ltd., Zhuzhou 412000, China
*
Author to whom correspondence should be addressed.
Actuators 2024, 13(10), 410; https://doi.org/10.3390/act13100410
Submission received: 8 September 2024 / Revised: 7 October 2024 / Accepted: 10 October 2024 / Published: 11 October 2024
(This article belongs to the Section Actuators for Land Transport)

Abstract

:
The incompleteness of the existing knowledge graphs in the railway domain creates information gaps, impacting their quality and effectiveness in the operation and maintenance of high-speed railway turnout switch machines. To address this, we propose a multi-layer model (KBGC) that combines KG-BERT, graph attention network (GAT), and Convolutional Embedding Network (ConvE) for knowledge graph completion in railway maintenance. KG-BERT fine-tunes a pre-trained BERT model to extract deep semantic features from entities and relationships, converting them into graph structures. GAT captures key structural relationships between nodes using an attention mechanism, producing enriched semantic and structural embeddings. Finally, ConvE reshapes and convolves these embeddings to learn complex entity interactions, enabling accurate link prediction. Extensive experiments on the HRTOM dataset, containing triplet data from high-speed railway turnout switch machines, demonstrate the model’s effectiveness, achieving an MRR of 50.8% and a Hits@10 of 60.7%. These findings show that the KBGC model significantly improves knowledge graph completion, aiding railway maintenance personnel in decision making and preventive maintenance, and providing new tools for railway maintenance applications.

1. Introduction

With the expansion of railway networks and increasing train speeds in China, the demand for reliable and safe signal infrastructure in the railway industry is rising [1]. There are many types of signal equipment failures, and their impact range is extensive [2]. As the actuator in the railway turnout switch system, the turnout switch machine converts electrical energy into mechanical energy through an internally mounted motor, driving the switch machine to control the switching of railway turnouts and enabling the train to switch and steer between different tracks. Its performance directly impacts the safety and efficiency of railway transportation [3], making it a key focus in high-speed railway signal equipment maintenance. It is crucial to detect all types of faults, as fault detection is instrumental in improving the level of intelligence of railway operations and optimizing their efficiency [4]. Numerous railway researchers have adopted data-driven approaches to analyze equipment failures. Reference [5] investigated the ZDJ9 switch conversion power curve, summarizing typical on-site failure types and causes, and developed a high-speed railway switch fault diagnosis model using the improved SSA-SVM algorithm. Reference [6] focused on the normal conversion and action curves of the S700 switch machine under failure conditions, using deep learning algorithms to extract image features, and achieving real-time fault diagnosis via the DCNN-SVM method. Reference [7] performed fault diagnosis using the sound of the switch machine during operation, addressing the limitation of traditional methods that rely solely on electrical signals without considering the physical characteristics of electromechanical systems. The abovementioned studies on switch equipment failures primarily focus on structured data. As high-speed railway operating mileage accumulates, substantial amounts of diverse data, including text, video, and images related to switch equipment operation and maintenance, have been collected, making unstructured data, particularly text data, the dominant component of big data. These text data contain critical information for ensuring railway safety. These switch operation and maintenance data are stored in an unstructured text format, serving as a knowledge base for future fault identification, diagnosis, and repair. However, the unstructured text data for high-speed turnout switch machine operations and maintenance are still recorded manually, resulting in issues such as ambiguity and irregularity in many fault records, challenges in computer processing, time-consuming manual analysis, and incomplete fault information mining [8]. When equipment faults occur, maintenance personnel must determine the cause based on specific conditions and relevant documents, resulting in relatively inefficient repairs. A knowledge graph is a knowledge base structured as “entity–relationship–entity” triplets, depicting concepts, entities, and their interrelationships in the physical world [9], thus effectively representing complex information. Knowledge graphs have been applied in fields such as power [10], aviation [11], and disaster management [12]. In the railway domain, Reference [13] organizes multi-source knowledge of high-speed trains to construct the schema layer of a high-speed train knowledge graph. Based on this conceptual schema, a data layer is developed to build domain-specific knowledge graphs for various stages of high-speed train operations. Reference [14] constructs a multi-domain integrated high-speed train maintainability knowledge graph, effectively integrating multi-source data to assist designers in train maintainability design and improve product quality. Reference [15] analyzes railway onboard fault maintenance logs and constructs a knowledge graph for railway onboard equipment faults, enhancing the knowledge discovery capability of fault logs and supporting intelligent maintenance. Reference [16] used semi-structured data from high-speed railway turnout equipment and applied a reinforcement learning-based approach to construct a knowledge graph for fault handling, thereby assisting maintenance personnel in decision making. However, the knowledge graphs generated through manual or semi-automated methods are often sparse and incomplete, leading to low-quality outputs with limited applicability. Research on knowledge graph completion in the railway domain is limited. Thus, filling the missing parts of the knowledge graph triplets is essential for enhancing the knowledge system, improving structural integrity, and providing a robust data foundation for decision support, preventive maintenance, and other railway maintenance applications.
Knowledge graph completion seeks to predict and fill missing elements in triplets [17], including entity and relation prediction. Entity prediction refers to identifying the missing head or tail entity type in a knowledge triplet, while relation prediction identifies the missing relationship type [18]. Reference [19] proposed the TransE model for knowledge graph completion based on translation distance. This model embeds entities and relationships into a vector space and uses vector addition to represent relationships, minimizing the distance between entities and their corresponding translations. Although simple, efficient, and easy to implement, the model performs poorly in handling complex relationships and does not fully consider semantic information [20]. Reference [21] proposed the RESCAL model, which is based on tensor decomposition. The model represents the knowledge graph as a high-order tensor and decomposes it to learn the low-dimensional embeddings of entities and relationships. However, it is susceptible to data sparsity [22]. Reference [23] introduced the ConvE model, a neural network-based approach that optimizes parameters through end-to-end training, capturing complex nonlinear relationships and higher-order features. However, the captured interaction relationships remain limited, focusing solely on the interaction information of stacked matrices, and the model fails to fully mine the semantic information between triplets [18]. Reference [24] proposed the KG-BERT knowledge completion model based on a pre-trained language model. This model leverages the pre-trained knowledge of the BERT module, which is highly effective in extracting textual semantic information and has demonstrated excellent performance in processing railway text data [25,26]. However, the model does not fully exploit the graph structure information of the knowledge graph, leading to suboptimal performance in knowledge graph completion. Although knowledge completion models based on graph neural networks (GNNs) utilize graph data representation that aligns with the inherent structure of knowledge graphs [27], GNNs achieve higher accuracy and greater robustness in graph-related tasks across various fields [28]. Nevertheless, the extraction of semantic information features remains insufficient [29]. The aforementioned models have limitations in handling complex relationships, semantic information, global graph structures, and local interaction features within knowledge graph triplets. Knowledge graph completion has been applied in various fields, including literature [30], agriculture [31], and disaster management [32]. In the railway sector, ensuring the completeness and accuracy of knowledge graphs is crucial for effective maintenance. Knowledge graph completion can fill gaps in the existing knowledge bases, enhance knowledge graph quality, and ensure that fault-related information is captured and utilized during the maintenance of core equipment like high-speed railway turnout switch machines. This significantly improves maintenance decision-making accuracy, reduces failure rates, and extends equipment lifespan.
In summary, a multi-level network model, referred to as the KBGC model, that integrates KG-BERT, GAT, and ConvE is proposed for the knowledge graph completion of high-speed railway turnout switch machine operation and maintenance, incorporating specialized railway terminology. The model first leverages the pre-trained KG-BERT to extract semantic information inherent in the knowledge triplet data, which includes complex railway terminology. Next, the GAT encoder employs an attention mechanism to process graph structure data, highlighting important nodes and effectively capturing dependencies between complex nodes. Finally, the ConvE decoder performs feature interactions, integrating semantic information with graph structure information to generate the optimal prediction value. A visual knowledge completion system for high-speed railway turnout switch machine operation and maintenance was developed by constructing a knowledge graph from the existing operation and maintenance data and applying the KBGC model for knowledge completion. This system supports on-site railway maintenance personnel in decision making and facilitates preventive maintenance.

2. Construction of High-Speed Rail Turnout Switch Machine Maintenance Knowledge Graph

During the construction of the high-speed railway turnout switch machine operation and maintenance knowledge graph, a combined top-down and bottom-up construction method is adopted to address the high level of specialization in the operational data. Natural language processing techniques, such as Chinese word segmentation and feature extraction, are employed to extract information from the text [33]. First, for entity extraction from unstructured data, the BERT-BiLSTM-CRF [34] model is introduced. By leveraging BERT’s robust text representation capabilities, text data are transformed into character-level embedding vectors. Contextual features in the text are captured using a bidirectional long short-term memory network (BiLSTM), while the conditional random field (CRF) layer learns internal dependencies between labels to achieve the optimal label sequence, ensuring accurate entity extraction. Next, relationship extraction is conducted using the BERT-CNN [35] model. BERT encodes the text data into word vectors, which are then fed into a convolutional neural network (CNN) for training. Finally, the model utilizes the fully connected and softmax layers to extract the semantic relationships between the target entities. For example, the text: “The automatic shutter of turnout J3 at #8 failed to disconnect the starting circuit at the first set of contacts after the turnout was in position, resulting in no indication for the turnout. It was restored after manually actuating the automatic shutter multiple times. The shutter was replaced during the maintenance window on the morning of the 30th. Following an on-site inspection by the manufacturer’s personnel, it was confirmed that the automatic shutter was stuck due to poor maintenance, which was classified as a signaling and communication responsibility”, extracts the entities “Automatic shutter of turnout J3 at #8”, “Failed to disconnect the starting circuit at the first set of contacts”, “No indication for the turnout”, “Manually actuating”, “Maintenance window”, “Replaced”, “Automatic shutter was stuck”, “Poor maintenance”, and the relationship “resulting in”. However, inconsistencies in recording methods for high-speed rail turnout switch machine maintenance data can result in varying descriptions of the same equipment, fault phenomena, and handling measures, causing duplicate entity extraction during knowledge graph construction. For instance, “turnout no indication” and “turnout lost indication” are named differently but represent the same equipment entity. To resolve this issue, a specialized dictionary is constructed based on the entities in the high-speed railway switch machine operation and maintenance data (as shown in Figure 1). The cosine similarity algorithm is then used to match the extracted entities with the dictionary, effectively linking them to their corresponding unique entities.
The processed knowledge triplet dataset, named the HRTOM dataset, is stored in the Neo4j graph database, completing the construction of the high-speed railway turnout switch machine maintenance knowledge graph. A partial visualization of this knowledge graph is shown in Figure 2.
The constructed knowledge graph consists of 891 entities, 15 relationships, and 10,027 triplets in total. To enhance the application of HRTOM triplet data in the operation and maintenance of high-speed railway turnout switch machines, the data were categorized into eight groups with the guidance of railway experts, as shown in Table 1.

3. High-Speed Railway Turnout Switch Machine Maintenance Knowledge Graph Completion Model

When completing the high-speed rail turnout switch machine maintenance knowledge graph, it is essential to consider both the semantic information and the graph structure of the maintenance data. Figure 3 illustrates the KBGC multi-level model for the knowledge graph completion of high-speed railway turnout switch machine operation and maintenance developed in this study. The model consists of three core modules—KG-BERT, GAT, and ConvE—designed to enhance knowledge graph completion accuracy through the deep processing of textual and structured information. First, the knowledge graph triplet (h, r, and t) is converted into a text sequence, and a fine-tuned KG-BERT model is applied to extract the deep semantic features of the entities and relationships within the operation and maintenance texts of high-speed railway turnout switch machines. Next, the graph attention network (GAT) layer is employed to process the relevance and importance of different nodes, effectively extracting structural information from the knowledge graph. Finally, the ConvE decoder module extracts higher-order interaction features through convolution operations, effectively completing the missing parts of the knowledge graph, which can subsequently be utilized for switch machine maintenance decision making.

3.1. KG-BERT Layer

The fine-tuned KG-BERT model utilizes BERT’s robust semantic extraction capabilities. In the high-speed railway turnout switch machine maintenance knowledge graph completion task, triplet data are first converted into text sequences and then transformed into vector representations via the BERT embedding layer, as shown in Figure 4. First, the triplet is converted into a text sequence with special tokens, where the [CLS] token is added at the beginning of the sequence and [SEP] is used to mark the end. Then, the pre-trained BERT model encodes the sequence, converting each token into token embedding, segment embedding, and position embedding, which are summed to form a combined embedding, represented by the following formula:
x i = e i + p i + s i
In the formula, x i represents the final embedding of the i-th word in the input sequence. Token embeddings decompose the text of the entities and relations into tokens using a tokenizer and convert them into vectors. Segment embeddings treat the head entity, relation, and tail entity as distinct segments to differentiate the logical parts of the input sequence. Positional embeddings provide positional information for each token in the sequence, enabling the model to deeply understand the semantic connections between the entities and relationships. Next, the combined embeddings are fed into a multi-layered Transformer encoder. The Transformer encoder utilizes a self-attention mechanism to represent each word in the input sequence as a weighted sum of all the other words. The formula is given as follows:
Attention ( Q , K , V ) = softmax ( Q K T d k ) V
In the formula, Q , K , and V denote the query, key, and value vectors, respectively, generated by d . Generate the contextual semantic representation for each token, extracting the semantic embedding vectors of the head entity, relationship, and tail entity ( Y h , Y r , and Y t ). Finally, the generated embeddings are mapped onto the graph structure, where the head entity and tail entity serve as nodes, and the relationship acts as the edge connecting them, forming the graph structure ( h h , h r , h t ).

3.2. GAT Layer

The GAT layer strengthens the associations between the entities in the graph structure derived from the KG-BERT layer using a graph attention mechanism. The structure of the model is illustrated in Figure 5. By calculating attention weights between different entities, the model dynamically adjusts the information flow, enabling it to consider not only local connection features but also to capture global information across the entire graph structure.
The core of GAT is the calculation of attention coefficients, which determine the importance of each node i to its neighboring node j in the graph. The initial representation of nodes and edges is obtained based on the adjacency matrix A and the node feature matrix h.
A = ( a i j ) + I N 0
In the formula, A is the adjacency matrix, representing the connection relationship between node i and node j, and III is the identity matrix, used to add self-connections in the adjacency matrix. There are N nodes, each with a feature dimension of F, and the set of node features is represented as follows:
h = { h 1 , h 2 , , h N } h i R F
To obtain an updated feature matrix, a weight-sharing parameter matrix W R F × F is first applied to each node, Then, self-attention mechanisms are executed for each node to calculate the corresponding attention coefficients, denoted as e i j , representing the importance of node j to node i:
e i j = α ( W h i , W h j )
The specific calculation of the attention score is as follows:
α ( W h i , W h j ) = LeakyReLU ( a T [ W h i | | W h j ] )
In the formula, a is the vector used to calculate the attention score, || represents the vector concatenation operation, and LeakyReLU is the activation function. Next, the attention coefficients of all the neighboring nodes are normalized using the Softmax function, as shown in the formula:
a i j = exp ( e i j ) k N ( i ) exp ( e i k )
Finally, the features of the neighboring nodes are aggregated to obtain the new feature representation of the central node i, as given by the formula:
h i l + 1 = σ ( j N i a i j h j l W )
In the formula, h i l + 1 and h i l represent the feature vectors of node i after and before the update, respectively. N i is the set of neighboring nodes of node i, W is the weight matrix, and σ is the nonlinear activation function.
The GAT layer dynamically captures and aggregates graph structural information, generating node embeddings enriched with structural details. These embeddings are then passed as input to the ConvE layer for further processing and feature extraction.

3.3. ConvE Layer

The ConvE model is one of the most widely used decoders for evaluating the validity of knowledge graph triplets [36]. As illustrated in Figure 6, the model effectively processes interaction features between entities and relations using convolutional and fully connected layers to achieve precise link prediction. The ConvE layer fuses the output features of the GAT layer with relation-specific embeddings, integrating entity and relation vectors through element-wise concatenation and reshaping them into a comprehensive feature matrix, which is subsequently fed into the convolutional layer. Convolutional filters capture specific relationships in the data from various perspectives, extracting key local features of entity–relation interactions. The convolved feature map is flattened and transformed into a one-dimensional vector, which is then integrated and projected into the embedding space through the fully connected layer. Next, matrix multiplication is performed on the entity matrix and relation matrix to compute the triplet score. The scoring function for the triplet is defined as follows:
p ( h , r , t ) = ReLU ( vec ( ReLU [ e h ; e r ] ω ) W ) ) e t
In the formula, ω represents a set of filters; “ ” is the convolution operator; vec ( · ) transforms the tensor into a vector; and ReLU is the activation function. Finally, the Logistic Sigmoid function is used to generate the prediction values for entities and relations.

3.4. Loss Function

To optimize the model’s predictive ability, a negative sampling loss function is used. For each positive sample triplet (e.g., “the switch detector’s movable contact is broken, measures are taken, the contact group is replaced”), several negative samples are generated by randomly replacing entities or relations. The loss function adjusts the model parameters by comparing the positive and negative samples, ensuring that the positive samples score higher than the negative samples, thereby improving prediction accuracy. The loss function is formulated as follows:
L = ( h , r , t ) S + lg ( σ ( f ( h , r , t ) ) ) + ( h , r , t ) lg ( σ ( f ( h , r , t ) ) )
In the formula, S + represents the set of positive samples, f ( h , r , t ) is the score of the model for the triplet ( h , r , t ) , and σ represents the Sigmoid function.

4. Empirical Results and Discussion

4.1. Experimental Datasets and Evaluation Metrics

The experimental data involved in this paper are the on-site high-speed railway turnout switch machine maintenance data from various railway bureau jurisdictions, covering nearly five years of high-speed railway turnout switch machine maintenance data. They include the component-level causes of turnout switch machine equipment malfunctions and failures, such as phase failure protectors, indication rod insulation, switch machine contact groups, turnout grounding leakage devices, and close-contact detectors. The HRTOM dataset is randomly divided into training, validation, and test sets in a ratio of 8:1:1 to conduct comparative experiments on the model. Mean Reciprocal Rank (MRR) [37] and Hits@k [38], commonly used in knowledge graph completion research, are selected as the evaluation metrics.
MRR = 1 | S | i = 1 | S | 1 r a n k i = 1 | S | 1 r a n k 1 + 1 r a n k 2 + + 1 r a n k | S |
In the formula, S represents the number of triplets, and r a n k i represents the predicted rank of the ith triplet. The value of MRR ranges from [0, 1], with values closer to 1 indicating more accurate model predictions.
Hits @ k = 1 | S | i = 1 S I ( r a n k i k )
In the formula, Hits@k refers to the average proportion of triplets ranked less than k in triplet prediction; I represents the indicator function.

4.2. Experimental Parameter Settings

The experimental environment consists of a CPU (Intel i9 12900K, Intel, Santa Clara, CA, USA), GPU (NVIDIA GeForce RTX3080Ti, NVIDIA Corporation, Santa Clara, CA, USA), and the operating platform is Pycharm2021.2. The code is written in Python 3.8 using the deep learning framework Pytorch. After multiple experiments, the main parameter settings for the model to achieve optimal performance are shown in Table 2.

4.3. Comparative Experiments

To validate the performance of the proposed model in the knowledge graph completion task for high-speed railway turnout switch machine operation and maintenance, comparative experiments were conducted under identical conditions using various models, including translational distance-based models TransE and TransH, tensor decomposition-based models Rescal and DistMult, a neural network-based model ConvE, and a graph neural network-based model KBGAT. The results are presented in Figure 7.
As shown in the figure, the proposed model outperforms the compared models in terms of MRR and Hits@10 on the HRTOM dataset, demonstrating the superiority of the KBGC model. Compared to the translation-based model TransH, the proposed model shows a 6.2% and 10.2% improvement in MRR and Hits@10, respectively. Although TransH addresses TransE’s limitations in handling many-to-many relationships, it still struggles to capture the nonlinear interactions between relationships. The proposed model utilizes the GAT layer’s attention mechanism to effectively identify and reinforce the dependencies between key nodes, better capturing complex relationships and improving prediction accuracy. Compared to the tensor decomposition-based model RESCAL, the proposed model shows a 4.5% and 9.3% increase in MRR and Hits@10, respectively. This is because in the HRTOM dataset, certain entities and relationships appear infrequently, and RESCAL struggles to handle sparse data, leading to poor prediction performance. By leveraging the KG-BERT layer’s ability to extract deep semantic features and the GAT layer’s processing of structural information, the proposed model maintains high prediction accuracy even with sparse data. Compared to the graph neural network-based model KBGAT, which also incorporates graph attention networks, the proposed model achieves better results, with MRR and Hits@10 increasing by 5.6% and 3.1%, respectively. This indicates that the proposed model can more comprehensively extract contextual information from nodes, enhance semantic feature representation in triplets, and achieve superior performance in complex data scenarios. Specifically, as shown in Figure 8, for the triplet (Automatic contact breaker in turnout switch machine, Occurs, Breakage of the actuating roller), the proposed model, in addition to identifying “Breakage of the actuating roller”, also predicts other possible fault phenomena for the automatic contact breaker in turnout switch machine, such as “Sticking or jamming of the pivot shaft”, “Insufficient tensile force of the tension spring”, “Fracture of the moving contact holder”, and “Short circuit and arcing of the stationary contact”. The TransH model completion results include “Sticking or jamming of the pivot shaft”, “No indication of positioning”, “Fracture of the moving contact holder”, and “Short circuit and arcing of the stationary contact”. The KBGAT model completion results include “Sticking or jamming of the pivot shaft”, “Insufficient tensile force of the tension spring”, “Fracture of the moving contact holder”, and “Midway stalling of the turnout switch machine”. Clearly, “No indication of positioning” and “Midway stalling of the turnout switch machine” are not direct faults of the automatic switch itself, but secondary fault phenomena caused by it. Therefore, it can be concluded that the proposed model outperforms the other models in terms of fault completion accuracy and rationality.

4.4. Ablation Experiments

To investigate the role of each component in the KBGC model architecture for high-speed railway turnout switch machine operation and maintenance knowledge graph completion, and to analyze the impact of these components on the model’s overall performance, several ablation experiments were conducted on the HRTOM dataset under identical experimental conditions. The results of these experiments are presented in Figure 9.
(1)
GAT + ConvE: The KG-BERT pre-trained model is excluded, leaving the GAT and ConvE modules. GAT captures the graph structure information by updating entity representations through graph edges (relationships). ConvE subsequently extracts local features from the entity and relationship embeddings via convolution operations, thereby enhancing the model’s predictive capability.
(2)
KG-BERT + GAT: The ConvE module is removed, retaining the KG-BERT and GAT modules. KG-BERT embeds the entities and relationships from the triplets into a vector space, utilizing the pre-trained BERT model for prediction to capture complex semantic information. While KG-BERT encodes the input, GAT updates the entity representation by considering the influence of the neighboring entities on the target entity. The attention weights are used to determine the importance of different neighbors, providing more comprehensive modeling and prediction of entities and relationships.
(3)
KG-BERT + ConvE: The GAT module is excluded, leaving the KG-BERT and ConvE modules. KG-BERT utilizes the BERT pre-trained model to capture global semantic information in the knowledge graph, while ConvE aggregates the local feature information of the entities and relationships through convolution operations. Although this combination balances semantic understanding and local feature extraction, it lacks supplementary graph structure information.
(4)
KBGC: As the complete model in the experiment, the above three models are integrated to form a multi-level architecture. KG-BERT performs the global semantic encoding of entities and relationships using the BERT model, capturing complex semantic information. The GAT module extracts graph structure features from the knowledge graph, supplementing local structural information. The ConvE convolutional network then aggregates features through interactions between the entities and relationships, thus accomplishing the completion task.
As illustrated in Figure 8, the proposed model outperforms the other models in terms of performance metrics. Compared to the GAT + ConvE model, the proposed model shows a 1.2% and 5.1% increase in MRR and Hits@10, respectively. This is because the GAT + ConvE model’s ability to represent semantic information is limited when extracting entities and relationships from triplets, resulting in lower prediction accuracy. This indicates that incorporating the KG-BERT model as a pre-training strategy effectively enhances the extraction of intrinsic triplet features, optimizes the training efficiency of the graph attention network, and improves the representation of entity and relationship encoding, thereby boosting link prediction performance. Compared to the KG-BERT + GAT model, the proposed model shows a 1.7% and 4.4% increase in MRR and Hits@10, respectively. This is because without ConvE, the KG-BERT + GAT model struggles to extract effective local features from long-tail entities, resulting in poor performance when dealing with rare entities. Consequently, the model’s ability to generalize to long-tail entities and relationships declines during prediction. Compared to the KG-BERT + ConvE model, the proposed model shows a 2.9% and 3.9% increase in MRR and Hits@10, respectively. This is because without GAT, the KG-BERT + ConvE model can only capture dependencies for single-hop or short-path relationships. It fails to model long-path dependencies (e.g., multi-hop reasoning) or cross-entity relationship patterns, resulting in reduced prediction accuracy. The proposed model leverages graph structure, global semantic information, and convolutional features, thus achieving superior performance in triplet prediction tasks.

4.5. Model Generalization Analysis

To evaluate the generalization ability of the proposed model, experiments were performed on the public datasets FB15k-237 [39] and WN18RR [23]. The experimental results are presented in Table 3. The results indicate that the proposed model achieves strong performance, with the MRR values on FB15k-237 and WN18RR improving by 1.8% and 4.5%, respectively, compared to the state-of-the-art KBGAT model. These findings demonstrate that the proposed KBGC model exhibits excellent generalization ability.

5. Knowledge Completion Visualization Application

The model predicts the five most relevant missing entities based on their associations with the current entity and relationship. For example, a “Poor contact of turnout close contact detector” fault may cause the phenomenon of “Indication circuit cannot be connected”. After model completion, it may also result in other faults such as “Idling”, “ Improper Switching of Turnout”, “Turnout obstruction”, and “No indication for turnout.” Figure 10 illustrates three types of completion examples: (Fault Phenomenon, Cause, Fault Phenomenon), (Equipment Phenomenon, Take, Maintenance Measure), and (Fault Location, Occur, Fault Phenomenon).
To assist maintenance personnel in effectively utilizing the completed missing information, this study developed a knowledge completion system for high-speed railway turnout switch machine maintenance using the Flask framework, as illustrated in Figure 11. By inputting an entity and relationship from the triplet, the system will automatically identify the top five missing pieces of information most relevant to the current input and rank them in descending order of relevance (displayed from top to bottom in the system interface). By completing and inferring missing information from the existing data, the system predicts equipment operating conditions, failure causes, safety hazards, and maintenance measures. This assists railway maintenance personnel in better understanding turnout failure mechanisms, supports decision making, enables preventive maintenance, and accelerates the development of intelligent maintenance for high-speed railway fault handling.

6. Conclusions and Outlook

A knowledge graph completion model based on the KBGC framework is proposed to improve the maintenance of high-speed railway turnout switch machines. The model leverages deep learning techniques to comprehensively extract feature information from knowledge graph triplets. Its superiority and generalization capabilities are validated through comparisons with other models and evaluations on public datasets. By improving the completeness of the knowledge graph for railway switch machine maintenance, the model significantly enhances the accuracy and reliability of graph data, thereby strengthening intelligent decision-making and information services. Additionally, a user-friendly visualization system for the knowledge graph completion process was developed using the Flask framework, offering an intuitive interface for human–computer interaction. This system facilitates the practical application of the knowledge graph by enabling maintenance personnel to quickly access relevant information for predictive maintenance, thereby gaining a deeper understanding of fault mechanisms in high-speed railway turnouts and assisting them in making informed decisions to achieve effective preventive maintenance. In future work, we plan to optimize the current model by developing a lightweight version to reduce training time and computational costs, facilitating its deployment in large-scale knowledge graphs. Moreover, the integration of interpretability modules, such as case-based reasoning and multi-level attention visualization tools, will be prioritized to make the system outputs more comprehensible for non-technical users, such as railway maintenance personnel. This will enable them to better comprehend the model-generated results and make informed decisions based on these insights. Additionally, a feedback mechanism will be integrated into the knowledge graph completion system, allowing users to provide feedback on the prediction results, thereby improving the model’s performance and enhancing its ability to support precise and effective decision making for maintenance tasks.

Author Contributions

Conceptualization, H.L. and J.B.; methodology, N.H. and Z.Z.; data curation, W.B. and D.L.; writing—original draft preparation, J.B.; writing—review and editing, H.L.; visualization, J.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Key Research and Development Program of Gansu Province-Industry (No. 23YFGA0046), and the Experimental Teaching Reform Project (No. 20240309006).

Data Availability Statement

The data that have been used are confidential.

Conflicts of Interest

The authors Zhengxiang Zhao and Wansheng Bai were employed by the company CRRC Zhuzhou Institute Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Zhao, P.; Wang, X.; Fu, M. Fault Diagnosis of Switch Machines with Small Samples Based on Bayesian Meta-Learning. J. Railw. Sci. Eng. 2023, 20, 4008–4020. [Google Scholar]
  2. Xie, M.; He, J.; Hu, X.; Cao, Y. Ground Signal Fault Diagnosis of Urban Rail Transit Based on Fault Logs. J. Beijing Jiaotong Univ. 2020, 44, 27–35. [Google Scholar]
  3. Hu, X.; Cao, Y.; Tang, T.; Sun, Y. Data-Driven Technology of Fault Diagnosis in Railway Point Machines: Review and Challenges. Transp. Saf. Environ. 2022, 4, tdac036. [Google Scholar] [CrossRef]
  4. Hu, X.; Tang, T.; Tan, L.; Zhang, H. Fault Detection for Point Machines: A Review, Challenges, and Perspectives. Actuators 2023, 12, 391. [Google Scholar] [CrossRef]
  5. He, H.; Dai, M.; Li, X.; Tao, W. Research on Intelligent Fault Diagnosis Method for Turnouts Based on DCNN-SVM. J. Railw. Soc. 2023, 45, 103–113. [Google Scholar]
  6. Wang, Y.; Mi, G.; Kong, D.; Yang, J.; Zhang, Y. Research on High-Speed Railway Switch Fault Diagnosis Method Based on MDS and Improved SSA-SVM. J. China Railw. Soc. 2024, 46, 81–90. [Google Scholar]
  7. Liang, X.; Dai, S. Research on Switch Machine Fault Diagnosis Based on Sound Signals [J/OL]. Railw. Stand. Des. 2023, 1–9. [Google Scholar] [CrossRef]
  8. Hu, X.; Niu, R.; Tang, T. Pre-Processing of Metro Signaling Equipment Fault Text Based on Fusion of Lexical Domain and Semantic Domain. J. China Railw. Soc. 2021, 43, 78–85. [Google Scholar]
  9. Zhu, Q.; Wang, S.; Ding, Y.; Zeng, H.; Zhang, L.; Guo, Y.; Li, H.; Wang, W.; Song, S.; Hao, R.; et al. Construction Method of Knowledge Graph for Safety, Quality, and Progress in Intelligent Management of Railway Tunnel Blasting Construction. J. Wuhan Univ. Inf. Sci. Ed. 2022, 47, 1155–1164. [Google Scholar] [CrossRef]
  10. Huang, J.; Yang, C.; Feng, T.; Sun, N.; Li, J.; Yu, H.; Cui, H. Research and Application of Knowledge Graph Construction for Wind Turbine Operation and Maintenance. Power Syst. Prot. Control 2024, 52, 167–177. [Google Scholar] [CrossRef]
  11. Chen, Y.; Liu, K.; Wang, S. Research on Knowledge Graph Construction for Aviation Equipment Faults Based on BiGRU-Attention Improvement. Acta Aeronaut. Astronaut. Sin. 2023, 42, 1–13. Available online: http://kns.cnki.net/kcms/detail/11.1929.V.20240104.1523.004.html (accessed on 3 June 2024).
  12. Shen, W.; Zhong, Y.; Wang, J.; Zheng, Z.; Ma, A. Construction and Application of a Knowledge Graph for Flood Disasters Based on Multimodal Data. J. Wuhan Univ. Inf. Sci. Ed. 2023, 48, 2009–2018. [Google Scholar] [CrossRef]
  13. Wang, S.; Li, X.; Li, R.; Zhang, H. Knowledge Fusion Method for High-Speed Trains Based on Knowledge Graphs. J. Southwest Jiaotong Univ. 2024; in press. [Google Scholar]
  14. Guo, H.; Li, R.; Zhang, H.; Wei, Y.; Dai, Y. Construction of a Knowledge Graph for High-Speed Train Maintainability Design Based on Multi-Domain Integration. Chin. Mech. Eng. 2022, 33, 3015–3023. [Google Scholar]
  15. Xue, L.; Yao, X.; Zheng, Q.; Wang, X. Research on the Construction Method of a Knowledge Graph for High-Speed Rail Control Onboard Equipment Faults. J. Railw. Sci. Eng. 2023, 20, 34–43. [Google Scholar] [CrossRef]
  16. Lin, H.; Zhao, Z.; Lu, R.; Bai, W.; Hu, N.; Lu, R. Construction and Application of a Knowledge Graph for Fault Handling of High-Speed Railway Turnout Equipment. J. Railw. Soc. 2024, 45, 73–80. [Google Scholar]
  17. Shen, T.; Zhang, F.; Cheng, J. A Comprehensive Overview of Knowledge Graph Completion. Knowl.-Based Syst. 2022, 255, 109597. [Google Scholar] [CrossRef]
  18. Du, X.; Liu, M.; Shen, L.; Peng, X. A Survey on Knowledge Graph Representation Learning Methods for Link Prediction. J. Softw. 2024, 35, 87–117. [Google Scholar] [CrossRef]
  19. Bordes, A.; Usunier, N.; Garcia-Duran, A.; Weston, J.; Yakhnenko, O. Translating Embeddings for Modeling Multi-Relational Data. Adv. Neural Inf. Process. Syst. 2013, 26, 2787–2795. [Google Scholar]
  20. Sa, R.; Li, Y.; Lin, M. A Review of Knowledge Graph Reasoning and Question Answering. J. Comput. Sci. Explor. 2022, 16, 1727–1741. [Google Scholar]
  21. Nickel, M.; Tresp, V.; Kriegel, H.P. Factorizing YAGO: Scalable Machine Learning for Linked Data. In Proceedings of the 21st International Conference on World Wide Web, Lyon, France, 16–20 April 2012; pp. 271–280. [Google Scholar]
  22. Lin, Y.K.; Han, X.; Xie, R.B.; Liu, Z.Y.; Sun, M.S. Knowledge Representation Learning: A Quantitative Review. arXiv 2018, arXiv:1812.10901. [Google Scholar]
  23. Dettmers, T.; Minervini, P.; Stenetorp, P.; Riedel, S. Convolutional 2D Knowledge Graph Embeddings. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; Volume 32. No. 1. [Google Scholar]
  24. Yao, L.; Mao, C.; Luo, Y. KG-BERT: BERT for Knowledge Graph Completion. arXiv 2019, arXiv:1909.03193. [Google Scholar]
  25. Lin, H.; Bai, W.; Zhao, Z.; Hu, N.; Li, D.; Lu, R. Knowledge Extraction Method for High-Speed Railway Switch Maintenance Text. J. Railw. Sci. Eng. 2024, 21, 2569–2580. [Google Scholar] [CrossRef]
  26. Zhang, Y.; Ye, H.; Zhang, L.; Xue, Y. BERT-Based Short Text Classification Model and Its Application in Fault Diagnosis of Railway CIR Equipment. Syst. Sci. Math. 2024, 44, 115–131. [Google Scholar]
  27. Sun, S.; Li, X.; Li, W.; Lei, D.; Li, S.; Yang, L.; Wu, Y. A Review of Graph Neural Network Applications in Knowledge Graph Reasoning. J. Comput. Sci. Explor. 2023, 17, 27–52. [Google Scholar]
  28. Wu, B.; Liang, X.; Zhang, S.; Xu, R. Advances and Applications of Graph Neural Networks. Chin. J. Comput. 2022, 45, 35–68. [Google Scholar]
  29. Zamini, M.; Reza, H.; Rabiei, M. A Review of Knowledge Graph Completion. Information 2022, 13, 396. [Google Scholar] [CrossRef]
  30. Gao, F.; Wang, J.; Duan, L.; Yue, K.; Li, Z. Knowledge Graph Completion of Folk Literature by Integrating Neighborhood Information. J. Beihang Univ. 2024; in press. [Google Scholar] [CrossRef]
  31. Ge, W.; Zhou, J.; Yuan, L.; Zheng, P. A Precision Fertilization Recommendation Model for Rice Based on Knowledge Graph and Case-Based Reasoning. Trans. Chin. Soc. Agric. Eng. 2023, 39, 126–133. [Google Scholar]
  32. Yang, Y.; Zhu, Y.; Chen, S.; Liu, X.; Li, H. Application of AI Chain with Collective Intelligence Strategy in Knowledge Reasoning for Dam Flood Control and Emergency Rescue. J. Hydraul. Eng. 2023, 54, 1122–1132. [Google Scholar] [CrossRef]
  33. Hu, X.; Niu, R.; Tao, T. Research on Entropy-Based Corrective Maintenance Difficulty Estimation of Metro Signaling. In Proceedings of the 2019 IEEE Intelligent Transportation Systems Conference (ITSC), Auckland, New Zealand, 27–30 October 2019; IEEE: New York, NY, USA, 2019; pp. 79–85. [Google Scholar]
  34. Liu, Y.; Wei, S.; Huang, H.; Lai, Q.; Li, M.; Guan, L. Named Entity Recognition of Citrus Pests and Diseases Based on the BERT-BiLSTM-CRF Model. Expert Syst. Appl. 2023, 234, 121103. [Google Scholar] [CrossRef]
  35. Liao, C.; Huang, Z.; Yang, J.; Shao, J.; Wang, T.; Lin, Y. Transformer Risk Assessment and Auxiliary Maintenance Decision-Making Method Based on Defect Text Recognition. High Volt. Eng. 2024, 50, 2931–2941. [Google Scholar] [CrossRef]
  36. Liu, D.; Fang, Q.; Zhang, X.; Hu, J.; Qian, S.; Xu, C. Knowledge Graph Completion Based on Graph Contrastive Attention Networks. J. Beihang Univ. 2022, 48, 1428–1435. [Google Scholar] [CrossRef]
  37. Pang, J.; Liu, X.; Gu, Y.; Wang, X.; Zhao, Y.; Zhang, X.; Yu, G. Knowledge Hypergraph Link Prediction Based on Multi-Granularity Attention Networks. J. Softw. 2023, 34, 1259–1276. [Google Scholar] [CrossRef]
  38. Chen, Z.; Wang, X.; Wang, C.; Zhang, S.; Yan, H. Time-Aware Knowledge Hypergraph Link Prediction. J. Softw. 2023, 34, 4533–4547. [Google Scholar] [CrossRef]
  39. Schlichtkrull, M.; Kipf, T.N.; Bloem, P.; Van Den Berg, R.; Titov, I.; Welling, M. Modeling Relational Data with Graph Convolutional Networks. In The Semantic Web: 15th International Conference on Semantic Web (ESWC), Heraklion, Crete, Greece, 3–7 June 2018; Springer International Publishing: Berlin/Heidelberg, Germany, 2018; pp. 593–607. [Google Scholar]
Figure 1. High-speed railway turnout switch machine operation and maintenance specialized dictionary.
Figure 1. High-speed railway turnout switch machine operation and maintenance specialized dictionary.
Actuators 13 00410 g001
Figure 2. High-speed rail turnout switch machine maintenance knowledge graph.
Figure 2. High-speed rail turnout switch machine maintenance knowledge graph.
Actuators 13 00410 g002
Figure 3. KBGC multi-level high-speed rail turnout switch machine maintenance knowledge graph completion model.
Figure 3. KBGC multi-level high-speed rail turnout switch machine maintenance knowledge graph completion model.
Actuators 13 00410 g003
Figure 4. BERT model based on high-speed rail turnout switch machine maintenance triplets.
Figure 4. BERT model based on high-speed rail turnout switch machine maintenance triplets.
Actuators 13 00410 g004
Figure 5. (a) The attention mechanism ( W h i , W h j ) utilized by our model is parameterized by a weight vector and applies a LeakyReLU activation function; (b) a depiction of multi-head attention (with K = 3 heads) performed by node 1 on its neighborhood. The distinct arrow styles and colors represent independent attention calculations. Features aggregated by each head are concatenated or averaged, resulting in h i l + 1 .
Figure 5. (a) The attention mechanism ( W h i , W h j ) utilized by our model is parameterized by a weight vector and applies a LeakyReLU activation function; (b) a depiction of multi-head attention (with K = 3 heads) performed by node 1 on its neighborhood. The distinct arrow styles and colors represent independent attention calculations. Features aggregated by each head are concatenated or averaged, resulting in h i l + 1 .
Actuators 13 00410 g005
Figure 6. ConvE Model.
Figure 6. ConvE Model.
Actuators 13 00410 g006
Figure 7. Model comparison chart.
Figure 7. Model comparison chart.
Actuators 13 00410 g007
Figure 8. Partial model completion comparison example.
Figure 8. Partial model completion comparison example.
Actuators 13 00410 g008
Figure 9. Ablation experiment results.
Figure 9. Ablation experiment results.
Actuators 13 00410 g009
Figure 10. Knowledge graph completion example.
Figure 10. Knowledge graph completion example.
Actuators 13 00410 g010
Figure 11. High-speed railway turnout switch machine maintenance knowledge completion system.
Figure 11. High-speed railway turnout switch machine maintenance knowledge completion system.
Actuators 13 00410 g011
Table 1. Triplet type.
Table 1. Triplet type.
Head Entity TypeRelationship TypeTail Entity TypeExample TripletsNumber of Triplets
Fault LocationAffiliationSwitch equipment(Switch locking device, affiliation, and external locking and installation)127
Fault PhenomenonCauseFault Phenomenon(Red light flashing during track switch, cause, and switch start circuit cutoff)565
Maintenance MeasureObtainMaintenance Result(Simulation experiment, obtain, and switch returns to normal)282
Maintenance ConditionPermitMaintenance Measure(Maintenance window, permit, and inspect the contact block assembly of the switch machine)1021
Maintenance MeasureVerifyMaintenance Experiment(Inspection and disconnection, verify, and simulation experiment)346
Equipment PhenomenonTakeMaintenance Measure(Irregular pointer, take, and joint repair by union departments)1932
Fault LocationOccurFault Phenomenon(Sliding plate, occur, and switch stock rail lacks lubrication)663
Fault PhenomenonDetermineFault Nature(Switch machine motor malfunction, determine, and poor maintenance)278
Table 2. Experimental parameter settings.
Table 2. Experimental parameter settings.
Parameter NameParameter Value
KG-BERT Learning Rate3 × 10−5
KG-BERT Embedding Dimension768
KG-BERT Batch Size16
GAT Dropout Rate0.3
Number of GAT Layers1
Number of GAT Attention Heads8
ConvE Convolution Kernel Size3
Optimization FunctionAdam
Table 3. Experimental results of the model on public datasets.
Table 3. Experimental results of the model on public datasets.
DatasetModelMRR/%Hits@1/%Hits@3/%Hits@10/%
FB15k-237TransE29.420.731.646.5
TransH27.119.830.344.2
RESCAL35.426.138.853.2
DistMult24.115.526.341.7
ConvKB25.316.027.942.2
KBGAT51.825.839.862.1
KBGC53.641.252.963.7
WN18RRTransE22.65.336.149.7
TransH45.99.839.351.7
RESCAL46.842.247.152
DistMult42.839.044.049
ConvKB24.75.436.452.4
KBGAT444548.658.1
KBGC48.543.348.059.3
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lin, H.; Bao, J.; Hu, N.; Zhao, Z.; Bai, W.; Li, D. Knowledge Graph Completion for High-Speed Railway Turnout Switch Machine Maintenance Based on the Multi-Level KBGC Model. Actuators 2024, 13, 410. https://doi.org/10.3390/act13100410

AMA Style

Lin H, Bao J, Hu N, Zhao Z, Bai W, Li D. Knowledge Graph Completion for High-Speed Railway Turnout Switch Machine Maintenance Based on the Multi-Level KBGC Model. Actuators. 2024; 13(10):410. https://doi.org/10.3390/act13100410

Chicago/Turabian Style

Lin, Haixiang, Jijin Bao, Nana Hu, Zhengxiang Zhao, Wansheng Bai, and Dong Li. 2024. "Knowledge Graph Completion for High-Speed Railway Turnout Switch Machine Maintenance Based on the Multi-Level KBGC Model" Actuators 13, no. 10: 410. https://doi.org/10.3390/act13100410

APA Style

Lin, H., Bao, J., Hu, N., Zhao, Z., Bai, W., & Li, D. (2024). Knowledge Graph Completion for High-Speed Railway Turnout Switch Machine Maintenance Based on the Multi-Level KBGC Model. Actuators, 13(10), 410. https://doi.org/10.3390/act13100410

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop