Next Article in Journal
Numerical Design and Analysis of Advanced Roof Systems in Architecture with Environmentally Friendly Low-Carbon Materials
Next Article in Special Issue
Memory-Enhanced Knowledge Reasoning with Reinforcement Learning
Previous Article in Journal
Properties and Durability of Cementitious Composites Incorporating Solid-Solid Phase Change Materials
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Commonsense-Guided Inductive Relation Prediction with Dual Attention Mechanism

Laboratory for Big Data and Decision, National University of Defense Technology, Changsha 410072, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2024, 14(5), 2044; https://doi.org/10.3390/app14052044
Submission received: 9 February 2024 / Revised: 25 February 2024 / Accepted: 26 February 2024 / Published: 29 February 2024
(This article belongs to the Special Issue Deep Learning for Graph Management and Analytics)

Abstract

:
The inductive relation prediction of knowledge graphs, as an important research topic, aims at predicting the missing relation between unknown entities with many real-world applications. Existing approaches toward this problem mostly use enclosing subgraphs to extract the features of target nodes to make predictions; however, there is a tendency to ignore the neighboring relations outside the enclosing subgraph, thus leading to inaccurate predictions. In addition, they also neglect the rich commonsense information that can help filter out less convincing results. In order to address the above issues, this paper proposes a commonsense-guided inductive relation prediction method with a dual attention mechanism called CNIA. Specifically, in addition to the enclosing subgraph, we added the multi-hop neighboring relations of target nodes, thereby forming a neighbor-enriched subgraph where the initial embeddings are generated. Next, we obtained the subgraph representations with a dual attention (i.e., edge-aware and relation-aware) mechanism, as well as the neighboring relational path embeddings. Then, we concatenated the two embeddings before feeding them into the supervised learning model. A commonsense re-ranking mechanism was introduced to filter the results that conformed to commonsense. Extensive experiments on WN18RR, FB15k-237, and NELL995 showed that CNIA achieves better prediction results when compared to the state-of-the-art models. The results suggested that our proposed model can be considered as an effective and state-of-the-art solution for inductive relation prediction.

1. Introduction

Knowledge graphs (KGs) are composed of organized knowledge in the form of factual triples (entity, relation, entity), and they form a collection of interrelated knowledge, thereby facilitating downstream tasks such as question answering [1], relation extraction [2], and recommendation systems [3]. However, even state-of-the-art KGs suffer from an issue of incompleteness [4,5], such as FreeBase [6] and WikiData [7]. To solve this issue, many studies have been proposed mining missing triples in KGs, in which the embedding-based methods become a dominant paradigm, such as TransE [8], ComplEx [9], RGCN [10], and CompGCN [11]. In particular, certain scholars have explored knowledge graph completion under low-data regime conditions [12]. In actuality, the aforementioned methods are often only suitable for transductive scenarios, which assumes that the set of entities in KGs is fixed.
However, KGs undergo continuous updates, whereby new entities and triples are incorporated to store additional factual knowledge, such as new users and products on e-commerce platforms. Predicting the relation links between new entities requires inductive reasoning capabilities, which implies that generality should be derived from existing datasets and extended to a broader spectrum of fields, as shown in Figure 1. The crux of the inductive relation prediction [13] resides in utilizing information that is not specifically tied to a particular entity. A representative strategy in inductive relation prediction techniques is rule mining [14], which extracts first-order logic rules from a given KG and employs weighted combinations of these rules for inference. Each rule can be regarded as a relational path, comprising a set of relations from the head entity to the tail entity, which signifies the presence of a target relationship between two entities. For example, consider the straightforward rule (X, part_of, Y) ∧ (Y, located_in, Z) → (X, lives_in, Z), which was derived from the KG depicted in Figure 1a. These relational paths exist in symbolic forms and are independent of particular entities, thus rendering them inductive and highly interpretable.
Motivated by graph neural networks (GNNs) that have the ability of aggregating local information, researchers have recently proposed GNN-based inductive models. GraIL [15] models the subgraphs of target triples to capture topologies. Based on GraIL, some works [16,17,18] have further utilized enclosing subgraphs for inductive prediction. Recent research has also considered few-shot settings for handling unseen entities [19,20]. SNRI [21] extracts the neighbor features of target node and path features, solves the problem of sparse subgraphs, and introduces mutual information (MI) maximization to model from a global perspective, which improves the prediction effect of inductive relationships.
Nonetheless, we still observe several issues from the existing literature: (1) Ignorance of neighboring information outside the enclosing subgraph. Inductive relation prediction models relying on the enclosing subgraph disregard the remaining neighbors of the node. However, neighbors in proximity to the target nodes harbor valuable information for inferring the relation; (2) Overlooking the influence of different neighboring relations. Existing methods overlook variations in the impact of distinct structures on subgraph modeling, as well as in the diverse relationships of target nodes regarding relation predicting; and (3) Generation of facts that violate commonsense. Some of the predictions generated violate commonsense, which can be easily avoided by introducing the commonsense knowledge for screening.
To address these issues, we propose a Commonsense-guided Neighboring relation InfoMax model based on a dual-Attention mechanism (CNIA). CNIA was built upon the popular inductive relation prediction framework but improves on the following aspects: (1) In order to fully leverage the neighboring relations of target nodes, we constructed a neighbor-enriched subgraph so as to include the useful neighboring information in addition to the enclosing subgraph; (2) To account for the structure of KGs, a dual attention mechanism was employed. This mechanism combined edge-aware attention, which addresses variations in the influence of different edges on target nodes, as well as relation-aware attention, which addresses the variations in the influence of diverse relationships on predicted relationships. Together, these two types of attention enable more accurate predictions within KGs. (3) A commonsense-based re-ranking strategy was used to update the score function of the triples to filter out less convincing prediction results. By integrating the aforementioned strategies during training, CNIA ensures the complete and effective retention of neighboring relation and path information, thereby enhancing the accuracy and rationality of KG relation predictions.
  • Contributions. The contributions of this work can be summarized into three aspects:
    • We put forward a commonsense-guided inductive relation prediction method with a dual attention mechanism, CNIA, which can enhance the representation learning and the accuracy of results;
    • We propose to construct a neighbor-enriched subgraph to retain more useful neighboring information to aid the prediction;
    • We compare the CNIA with state-of-the-art models on benchmark datasets, and the results demonstrate the superior performance of our model.
  • Organization. Section 2 provides an overview of the related works. Section 3 introduces the module. Section 4 describes the experimental design and analyzes the results, and this is followed by the conclusions in Section 5.

2. Related Works

In this section, we first present the existing solutions to inductive relation predictions. Next, we briefly introduce commonsense knowledge.

2.1. Relation Prediction Methods

In order to increase the completeness of KGs, state-of-the-art methods use a KG internal structure [8,15] or external KG information [22,23]. In this work, we focused on the former and studied the relation prediction problem.
Transductive methods. Transductive methods are used to learn an entity-specific embedding for each node, which have one thing in common: reasoning over original KGs. However, it is difficult to predict the missing links between unseen nodes. For example, TransE [8] is based on translation while RGCN [10] and CompGCN [11] are based on GNN. The major differences between them are the scoring function and whether the structure information in KGs is utilized. Wang et al. proposed to employ a global neighborhood aggregator to obtain the global structural information of an entity to solve the problem of sparse local structural information under certain snapshots [24]. Meng et al. proposed a multi-hop path inference model based on sparse temporal knowledge graphs [25]. Recently, Wang et al. proposed knowledge graph completion for multi-level interactions, where entities and relations interact simultaneously at both fine- and coarse-grain levels [26].
Inductive methods. Inductive methods can be used to learn how to reason on unseen nodes. There are two main categories of methods: rule-based and graph-based. Rule-based methods aim to learn entity-independent logical rules of reasoning. For instance, NeuralLP [14] and DRUM [27] integrate neural networks with symbolic rules to learn logic rules and rule confidence in an end-to-end microscopic manner.
Concerning graph-based methods, in recent years, researchers have drawn inspiration from the local information aggregation capability of graph neural networks, and they have incorporated graph neural networks (GNNs) into their models. GraIL [15] acquires the topology of target nodes by extracting the enclosing subgraph of the target triad, thus exhibiting inductive prediction abilities. Building on this model, TACT [16] introduces the correlation of relations in subgraphs and constructs a relational correlation network (RCN) to enhance the encoding of subgraphs. CoMPLIE [17] proposes a node-edge communication message propagation network to enhance the interaction between nodes and edges, and it naturally deals with asymmetric or anti-symmetric relations so as to enhance the adequate flow of relational information. ConGLR [13] formulates a contextual graph that represents the relationship paths in the subgraph, whereby two GCNs are applied to deal with the enclosing subgraphs and contexts, in which different layers utilize counterpart outputs interactively for better feature representation. RE-PORT [28] aggregates relational paths and contexts to capture the connections and intrinsic nature of entities through a unified hierarchical transformation framework. However, since the experimental metrics of RE-PORT are different from those of other state-of-the-art models, the RE-PORT model was not chosen as a comparison model in the experimental part of this paper. RMPI [29] uses a novel relational message passing network for fully inductive knowledge graph completion. SNRI [21] extracts the neighbor relational features and path embeddings of nodes to make full use of the full neighbor relational information of the entities to obtain a better inductive effect. However, all these methods only add additional simple processing load, and they do not take full advantage of the whole structural feature of KGs. Different from SNRI, CNIA keeps integrated neighboring relations, utilizes a dual attention mechanism to process the structural feature of subgraphs, and it introduces commonsense re-ranking.

2.2. Commonsense Knowledge

Commonsense knowledge is a crucial link in addressing the bottlenecks of AI and knowledge engineering technology. And the acquisition of commonsense knowledge is the basic problem in this field. The earliest construction method involves experts manually defining the schema and relationship types of the knowledge base. Lenat [30] constructed CYC, one of the oldest knowledge bases, in the 1980s. However, the expert construction method requires substantial human and material resources. Consequently, researchers have began developing semi-structured and unstructured text extraction methods. The commonsense knowledge base constructed by YAGO [31] contains more than 1 million entities and 5 million facts, which were derived from semi-structured Wikipedia data and harmonized with WordNet through a well-designed combination of rule-based and heuristic approaches.
The aforementioned approaches prioritize encyclopedic knowledge and structure storage by establishing a well-defined entity space and a corresponding relationship system. However, actual general knowledge is much more loosely structured and challenging to apply to the model of two entities with known relationships. Therefore, the existing solution is to model the entity parts as natural language phrases and the relations as any concepts that can connect the entities. The OpenIE approach, for example, reveals the properties of open-text entities and relations. However, the method is extractive and it is difficult to obtain the semantic information of the text.

3. Methodology

In this section, we introduce the framework of CNIA. Next, we describe the foundation framework of the inductive relation prediction. Finally, we elaborate on the components of CNIA. An overview of our proposed model, CNIA, is shown in Figure 2.

3.1. Problem Statement

The inductive relationship prediction aims to predict the relationships between entities outside the training set. Given the KG dataset and the target triple as inputs, the model needs to generate scores of the predicted relationships of the target triple as the outputs. Previous solutions toward this task fail to take into account the full adjacency features and commonsense constraints. To fill in these gaps, our proposed model CNIA extracts more comprehensive neighboring relational features. In addition, we applied a dual attention mechanism to improve the representation learning. We also adopt commonsense re-ranking, for the first time to our knowledge, to refine the prediction results.

3.2. Model Overview

The core idea of CNIA is that the subgraph structure and relationship information around two nodes can predict the relationship r t between two nodes u , v . Thus, only the structure and relationship information is extracted as the model input. Firstly, we extracted the neighboring relational features from the multi-hop neighbor subgraph surrounding the target nodes, which resulted in the creation of a neighbor-enhanced subgraph. We then utilized the subgraph neural network (SNN) to obtain the local subgraph embeddings from this enhanced subgraph. To capture the distinct influences of the different edges on target node embeddings and the different relationships associated with the target relations, we introduced edge-aware and relation-aware attention mechanisms. These mechanisms enabled us to learn more informative representations. Additionally, we modeled the neighboring relational paths to obtain relational path embeddings, which further contribute to the overall representation learning process. Subsequently, we concatenated the subgraph embeddings and relational path embeddings before inputting them into the supervised learning phase. And we utilized MI maximization to contrast the learning from global and local perspectives. Furthermore, we introduced a commonsense re-ranking mechanism to prioritize the results that align with commonsense knowledge and conform to general expectations. This mechanism ensures the selection of more reliable and sensible predictions. Figure 2 shows the steps of the whole process.
The notation table can be found in Table 1.

3.3. Foundation Framework

In this subsection, we introduce the general model of inductive relation predictions.
  • Neighboring Relational Feature Extraction. The neighboring relational feature model consists of two parts: subgraph extraction and node initialization. The local graph neighborhood of a specific triple in the KGs contained the logical evidence needed to infer the relationship between the target nodes; as such, the first step was to extract the subgraphs around the target nodes that contain complete neighbor relationships to initialize the node features. For the subgraph extraction, the target triple u , r t , v was identified and the enclosing subgraph G u , r t , v around the target triple was extracted. More details can be found in the GraIL paper [15]. Then, regarding node initialization—considering that the inductive inference method cannot utilize the node attribute features—the node initial features were obtained by extracting the node’s positional features and adjacency features. The details can be found in [21].
  • Subgraph Representaton Learning. As the main component of the foundation framework, this stage includes two parts: (1) obtaining the representation of subgraph entity nodes by the subgraph neural networks; (2) extracting and modeling the neighboring relational paths across target triples.
For subgraph modeling, the subgraph G u , r t , v of the target triad ( u , r t , v ) was first input into the GNN. SNRI defines the update strategy of the node features at each layer of the neural network. Inspired by CoMPILE [17], it feeds all node embeddings H L of the last layer to a gated recurrent unit to increase the expressive ability of the network. Finally, to obtain the representation of subgraph G, it uses an average readout function as follows:
h G = 1 V G i V G h i L .
The neighboring relational path modeling aims to obtain valuable information from the relationship paths of ( u , v ) as this helps to predict the type of relationship between u and v. The specific procedure can be found in reference [21].
The subgraph representation h G and neighboring relational path representation p G were concatenated as the final representation of subgraph s G as follows:
s G = h G p G .
The score function of target triple is
S c o r e u , r t , v = W s h u L h u L e r t L s G ,
where h u L , h u L , and e r t L represent the embeddings of the target nodes u, v, and target relation r t at the L layer of the SNN.
  • Supervised training. A loss function for supervised learning was constructed:
    L sup = u , r t , v G max 0 , S c o r e u , r t , v S c o r e u , r t , v + θ ,
    where u , r t , v and u , r t , v refer to the positive and negative samples, and  θ is the margin hyperparameter.
  • Contrastive Learning. Contrastive learning was widely used in unsupervised learning to attract positive samples and repel negative samples to obtain a higher-quality representation. To avoid the SNN in SNRI over-emphasizing the local structure, we further modeled the neighboring relations in a global way by subgraph–graph mutual information maximization, that is, SNRI sought to enable the neighboring relational features and paths to capture the global information of the entire KG, which would be realized in [21]. And the loss function L M I could also be obtained.
  • Joint Training. The ultimate learning goal of this task was to train the following loss function:
    L = L sup + λ L M I ,
    where λ controls the contribution of the MI max mechanism.

3.4. Our Model

Next, we introduce our model (cf. Figure 2), which improves over the foundation framework in four respects, i.e., neighbor-enriched subgraph extraction, neighboring relational paths, subgraph modeling based on a dual attention mechanism, and commonsense re-ranking.

3.4.1. Neighbor-Enriched Subgraph Extraction

Based on the enclosing subgraph extraction, the other neighboring relationships of each node were expanded to obtain the neighbor-enriched subgraph. The specific steps are as follows:
Step 1
Obtain the set of three-hop neighbor nodes N k u and N k v of the target nodes u and v in the KG, respectively. The neighbor nodes obtained here do not distinguish the direction.
Step 2
Take the intersection N k u N k v of the neighbor nodes of u and v to obtain the nodes of the enclosing subgraph.
Step 3
Filter out the isolated nodes or nodes with a distance that is greater than three from any of the target nodes to obtain the enclosing subgraph with a path length that does not exceed the length between the target nodes.
Step 4
Keep the complete three-hop neighbor relationship N r i of each node i, which includes the part omitted by the enclosing subgraph.

3.4.2. Neighboring Relational Path

As shown in Figure 3, to solve the sparse subgraph problem, we modeled the neighboring relational paths.
A relational path is a sequence of relations on a target node, e.g.,  p 1  = ( r 1 , r 2 ), where r 1 and r 2 are the surrounding relations on the target nodes u and v. Define P as the set of all neighboring relational paths on ( u , v ) in the subgraph. For each relation path p, GRU modeling is first applied as follows:
m u k = i N u h i k ,
h i k + 1 = σ m u k m v k h i k · W k + b k ,
where i N u is a neighboring edge of node u and h i 0 is the initial characterization of i. Repeat the above equation for K rounds, and then splice to obtain the path representation of the subgraph as follows:
p G = σ m u K 1 m v K 1 · W k 1 + b k 1 .

3.4.3. Subgraph Modeling Based on a Dual Attention Mechanism

For subgraph modeling, in order to improve the accuracy of the subgraph information utilization and distinguish the role played by different structures in modeling, we propose to integrate edge and relation dual attention mechanisms into a GCN. Generally, the entity embedding for updating subgraphs employs the traditional iterative message passing strategy of a GCN.
h v k + 1 = u , r N s j α u , r , v k W t 1 k ϕ h u k , e r k + W t 2 k h v k ,
where u , r , v is the example triple and N s · is the set of neighbors containing entity relationship pairs. W t 1 k and W t 2 k are transformation matrices, and h v k represents the embedding of entity v at layer k.
In actuality, there are differences between the neighbors in entity modeling. For example, neighbors ( h 3 , r 2 ) , as well as ( h 4 , r 6 ) , have different effects on the representation of h 2 . Meanwhile, predicting different target relationships affects the representation of h 2 . In order to distinguish these effects, we propose two types of attention, edge-aware β u , r k and relation-aware γ r , r t k , to compute the joint neighbor attention values.
β u , r k = σ W 1 k h u k e r k h v k ,
γ r , r t k = σ W 2 k e r k e r t k ,
α u , r , v k = softmax β u , r k + γ r , r t k ,
where W 1 k and W 2 k are transformation matrices.
To enable information sharing and improve representation, we integrated updated entity embeddings to enhance the representation as follows:
H i n d e x v k + 1 = λ 1 H i n d e x v k + 1 λ 1 h v k + 1 ,
where i n d e x v indicates the actual index of entity v.
Based on this, we can obtain the embedding matrix H L , which consists of the embedding vectors of all the entities in the subgraph of the last GCN. To further improve the representational power of the model, a two-way gated recursive unit GRU, as inspired by [17,32], was added after the last layer of the GCN. Since the entities in the subgraph are sorted according to the distances to the target entities, the GRU that handles the sequences can be used to increase the inter-entity interactions, thereby updating the embedding representation as follows:
H L = BiGRU H L .
Finally, the average readout function (1) was used to process H L to obtain the representation h G of the subgraph G.

3.4.4. Commonsense Re-Ranking

To ensure that the generated triples conform to commonsense, they are compared to commonsense knowledge. The generation of commonsense knowledge is specifically divided into four steps:
  • Fine tuning BERT to acquire contextual representations. The pre-trained language model BERT is suitable for acquiring the contextual representations of triples. Specifically, given a tuple ( u , r , v ) , u, r, and v are iteratively masked such that the encoder predicts the masked elements using the other two unmasked elements, which enables the BERT model to better understand the relationships between the tuple elements.
  • Filtering abstract concepts. Replacing ternary entities with entity concepts may result in concepts with a high level of abstraction. We use entity and concept representations to compute the probability of a concept appearing in an entity to measure the abstraction level of the concept. A threshold is set to filter out concepts with a higher abstraction level whose occurrence expectation is lower than the threshold.
  • Entity-to-concept mapping. After filtering out the higher degree of abstraction, concepts were used instead of entities in the triad to obtain conceptual triples. Commonsense knowledge in the individual form C1 was obtained by eliminating duplicate concept-level triples. Commonsense knowledge in the set form C2 was then obtained by merging concept-level triples that contained the same relations.
  • Filtering relationship-independent concepts. The commonsense knowledge obtained by substituting concepts for entities would also suffer from the problem that concept c is not related to relation e r . To measure the degree of relevance of concept c and relation e r , the following cosine similarity was calculated to obtain the similarity score.
    R r , c = cos e r , c ,
    e r = mean e i N r e i ,
    where N r denotes the set of relations r corresponding to the head entity or the tail entity, and  c and e r are the contextual representations of concepts and entities, respectively. A threshold was set to filter the target of relationship-independent concepts.
Based on the constructed commonsense knowledge database, we propose a simple commonsense-based re-ranking strategy, i.e., one that is based on the KGs of the train set or test set. Moreover, if the triple u , r t , v satisfies the commonsense c h , r t , c t , the score of the ternary is increased by ω > 0 .
S c o r e u , r t , v = W s h u L h v L e r t L s G + ω I u , r t , v c h , r t , c t ,
where I · is the indicator function. We substituted the updated triple score function into the original loss function (4). Then, we used MI maximization to obtain a global representation and train joint training strategy (5).

3.4.5. Algorithmic Descriptions

The algorithm of the whole model is presented in Algorithm 1.

4. Experiments

In this section, we first introduce the experimental configurations. Next, we provide the empirical results and discuss the performance, and then we finally discuss the further experiments that were conducted.

4.1. Experimental Configurations

In this subsection, we introduce the experimental configurations, including the datasets, evaluation metrics, parameter settings, and the baseline models.

4.1.1. Datasets

WN18RR [33], FB18k-237 [34], and NELL-995 [35] are widely used in transductive link prediction. As for inductive relation prediction, we follow GraIL to conduct experiments based on the variants of WN18RR, FB18k-237, and NELL-995, where entities in a test set are not contained in a train set. And every dataset induces four versions with increasing sizes.
Algorithm 1 The inductive process of the CNIA model
Input:  K G , target triple u , r t , v
Output: the S c o r e of u , r t , v
  1:
Extract subgraph G of target triple u , r t , v from K G
  2:
Initialize node representation h for all the nodes in G
  3:
for all k in K of GCN layers do
  4:
   Update the embedding h j k of each entity j at layer k
  5:
   Calculate edge attention value β r , r t k according to Equation (10)
  6:
   Calculate relation attention value γ r , r t k according to Equation (11)
  7:
   Generate general attention α u , r , v k according to Equation (12)
  8:
   Update the embedding matrix H index ( v ) k of all entities
  9:
end for
10:
Obtain entity embedding H L through BiGRU
11:
Obtain h G through a readout function
12:
Establish a Neighboring Relational Path Model to obtain path representation p G Equation (8)
13:
Establish score function S c o r e and loss function L s u p by Equations (3) and (4)
14:
Update S c o r e by commonsense re-ranking according to Equation (17)
15:
return  S c o r e
The specific sampling method was divided into the following three steps: (1) Uniformly sample several root entities as root nodes in the original KGs; (2) Take the concatenated set of the multi-hop neighbors of the root node as the training KGs. To prevent the number of nodes from growing exponentially, the size of k needs to be limited; (3) Remove the sampled training set from the entire original KGs and then repeat Steps 1-2. Then, sample the test set so as to ensure that the entities in the test set are not included in the training set. The detailed statistics of the dataset are shown in Table 2.

4.1.2. Evaluation Metrics

The majority of existing models, such as SNRI, use AUC-PR and Hits@10 as evaluation metrics. In order to make a fair comparison, we followed previous works and also used these two metrics. Specifically, AUC-PR is a classification indicator that computes the area under the precision–recall curve. Hits@10 is defined as the proportion of correct entities ranked in a top 10 of candidate entities. The results are obtained by averaging over five runs for an accurate evaluation.

4.1.3. Parameter Settings

In the subgraph extraction section, the experiment selected k = 3, indicating that only the neighboring nodes within three hops of the target node were retained for the union operation. The learning rate was set to 0.0005, the dropout rate was 0.5, and the embedding dimension was 32. The margin parameter θ for the supervised learning loss function was 10, and the coefficient λ for the joint loss function was 5. Simultaneously, L2 regularization was applied to prevent overfitting, the maximum training epoch was set to 50, and Adam was employed as the optimizer for the training model parameters. All experiments were conducted in PyTorch and executed on an NVIDIA RTX 3090.

4.1.4. Baseline Models

To explore the performance of our proposed model, we compared five state-of-the-art GCN-based baselines, including the following:
  • GraIL [15]: This method pioneered a novel approach to inductive reasoning by introducing subgraph encoding for the first time, whereby it addressed the invisible entities in entirely new KGs.
  • TACT [16]: This approach models the semantics between relationships and uses relationship topology to detect correlations for inductive relation predictions.
  • CoMPILE [17]: This method is grounded in the structure of local directed subgraphs, and it exhibits a robust inductive bias for handling entity-independent semantic relations.
  • SNRI [21]: This approach leverages the full adjacency of entities in subgraphs using neighbor relationship features and neighbor relationship paths. This forms the basis for inductive relation predictions in KGs.
  • ConGLR [13]: This method constructs a contextual graph that represents relational paths and subsequently processes them with two GCNs, each incorporating enclosing subgraphs.
  • RMPI [29]: This approach passes messages directly between relations to make full use of the relation patterns for subgraph reasoning.

4.2. Experimental Results

In this subsection, we first report the main results. Then, we discuss the ablation study that was subsequent conducted.

4.2.1. Main Results

The main experimental results are presented in Table 3 and Table 4. “Avg.” denotes the average metric values of the four versions of the same KG. The optimal and suboptimal results are marked in bold and underlined text, respectively. “—” represents the results that were not provided in the original work (and cannot be reproduced).
The two tables present the results of the comparative experiments. CNIA demonstrated optimal performance in both AUC-PR and Hits@10 metrics across most datasets, thus confirming the effectiveness and sophistication of CNIA. Specifically, the CNIA model outperformed the GraIL, TACT, and CoMPILE models on all datasets, and it surpassed the SNRI and ConGLR models on most datasets, albeit with suboptimal results on individual datasets. When averaging the metrics across the three datasets, it was evident that the average AUC-PR value of CNIA was optimal on all three datasets, while the Hits@10 value was suboptimal on WN18RR and optimal on the other two datasets.
Compared with the base model SNRI, the experimental results on both WN18RR and FB15k-37 were significantly improved, and the average AUC-PR and Hits@10 values were greatly increased, thereby indicating that CNIA is superior and can enhance the performance of inductive link prediction. Compared with the state-of-the-art model RMPI, CNIA is constantly better on the majority of datasets.
In comparison to the ConGLR model, the average AUC-PR improved by 0.44% on the WN18RR dataset, 0.25% on the FB15k-237 dataset, and 2.25% on the NELL-95 dataset. However, the average Hits@10 was lower than ConGLR on the WN18RR dataset, but it improved by 1.95% on the FB15k-237 dataset and 0.99% on the NELL-95 dataset. ConGLR only performed well on WN18RR-v2,v4 as on other datasets it did not perform well, which indicates that ConGLR’s biased logical reasoning method is specifically suited for certain WN18RR datasets. In contrast, our model CNIA consistently performed well across various datasets, thus demonstrating its versatility and stability across different scenarios.
Among the AUC-PR metric values, CNIA exhibited more pronounced advantages on the WN18RR and NELL-995 datasets. Upon observing Table 2, it was noted that the #R value of WN18RR and NELL-995 datasets consistently remained lower than that of FB15k-237, thus indicating that the subgraphs in these two datasets were more likely to be sparse and lack the structural information for reasoning. CNIA effectively leveraged the adjacent relationship information when compared to the baseline model, thereby addressing the issue of subgraph sparsity. However, the improvement effect on the FB15-237 dataset was not as noticeable, possibly because the subgraph density of this dataset was high and due to the fact that the effective structural information could be extracted using closed subgraphs, thus making the baseline model easily predictable.

4.2.2. Ablation Study

The ablation study explored the impact of each component on the overall model performance, and we conducted the ablation experiment using the WN18RR dataset. This experiment mainly consisted of five parts: (1) removal of the dual attention mechanism (w/o Att), (2) removal of the neighboring relational feature module (w/o NRF), (3) removal of the neighboring relational path module (w/o NRP), (4) removal of the commonsense-based re-ranking module (w/o CSR), (5) removal of the comparative learning (w/o CL). Table 5 shows the results of the ablation experiment, and we were able to find that the CNIA removal of any of the modules was worse than the original model, thus proving the effectiveness of each module.
  • CNIA w/o Att: Remove the dual attention mechanism and directly splice to obtain the neighboring relationship features. The Hits@10 was reduced by 3.72%, 1.25%, 5.37%, and 0.63%. This showed that ignoring the effect of different edges on nodes and the effect of different relations on the target relation reduces the accuracy of inference.
  • CNIA w/o NRF: Remove the neighbor relational features and predict directly from the node features obtained by initialization. Ignoring the neighbor relationship makes the node features less expressive, which results in losing the effective information, and it also cannot completely portray the nodes.
  • CNIA w/o NRP: Remove the neighboring relational path feature and ignore the message propagation path feature from the head node to the tail node. This makes the performance degradation obvious, thereby indicating that the neighboring relational path feature plays an important role in dealing with sparse subgraphs.
  • CNIA w/o CSR: Remove the commonsense re-ranking module that fails to filter out the predictions that do not conform to commonsense. The experiments verified that the absence of the commonsense re-ranking will produce relationships that do not conform to commonsense, thus degrading model performance.
  • CNIA w/o CL: Remove the contrastive learning and do not perform this operation of MI maximization. This resulted in the Hits@10 being reduced by 4.24%, 4.99%, 7.57%, and 9.24%. The results showed that, without contrastive learning, the results of most of the metrics dropped, thus demonstrating that global information helps to better model neighboring relational features.

4.2.3. Hyper-Parameter λ Analysis

In CNIA, λ governs the contribution of the contrastive learning MI InfoMax mechanism (Equation (5)). In this section, we further analyze the multiple values of lambda in [0, 0.5, 1.0, 5.0, 10.0]. The changes in Hits@10 values in WN18RR-v1, v2, v3, and v4 are depicted in Figure 4. It was observed that inductive relation prediction performance varies with λ values. Both excessively large and small λ values lead to a decrease in model performance.
When λ = 1 , the results are the worst on the v2 and v3 datasets, and they are moderate on the v1 and v4 datasets, thus indicating that the prediction performance is unstable. Thus, simply setting the weights of the two to be equal is not conducive to maximizing the performance. Notably, the effect is generally better when λ > 1 compared to λ < 1 , thereby indicating that contrastive learning plays a more significant role than supervised learning.
In summary, this discussion establishes that a better performance can be achieved when the coefficient λ is set to five, and it underscored the importance of contrastive learning through the ablation experiments.

4.2.4. Learning Rate Analysis

The learning rate is an important hyperparameter for controlling parameter updates during neural network training. The value of the learning rate affects the training efficiency. When the learning rate is too small, the learning speed is slow, which means that it is vulnerable to overfitting and the convergence speed is slow; when the learning rate is too large, the learning speed is fast but prone to the oscillation problem such that the training value fluctuates up and down in the optimal value.
The learning rate of the basic model SNRI in the experiment was set to 0.001. This was chosen in order to improve the efficiency and effect of the experiment. Next, we discussed the learning rate of the experiment and, respectively, took the values of 0.01, 0.05, 0.001, 0.005, and 0.0001 in the WN18RR dataset to conduct the experiment.
As shown in Figure 5, the experimental results on the v1 and v2 datasets were more stable when the learning rate was changed, with the Hits@10 remaining above 80%. When the learning rate was 0.01, the experimental effect was the worst on the v3 and v4 datasets. And when the learning rate was set to 0.0001, the Hits@10 decreased on all sub-datasets, thus indicating that the learning rate was too small to affect the prediction effect. The overall trend peaked at a learning rate of 0.005 and then fell back.

5. Conclusions

In this study, we proposed a novel inductive relation prediction method called CNIA, which leverages commonsense knowledge and employs a dual attention mechanism to enhance relationship inference accuracy. By fully utilizing neighbor relational information and incorporating a commonsense filtering process, CNIA was able to extract neighbor relational features through an edge-aware and relation-aware dual attention mechanism. Moreover, we employed contrastive learning for global modeling to further improve the performance of our model. The experimental results on the three benchmark datasets showed that CNIA outperforms other state-of-the-art models (we conducted an ablation study to verify the effectiveness of each module).
Regarding future research prospects, reasoning with multimodal knowledge graphs could be explored, where elements such as images can be added to improve inductive reasoning. KG completion with temporal constraints is also worthy of exploration, where temporal information should be taken into account when scoring triples. Building on existing research, we plan to extend our exploration into the temporal dimension and investigate inductive relation prediction methods on time series knowledge graphs. In addition, as large language models continue to advance, there is great potential to leverage these models to enhance inductive reasoning and build more accurate inductive models.

Author Contributions

Conceptualization, Y.D. and C.L.; methodology, W.Z. and Y.D.; software, H.X.; validation, J.T., H.X. and W.Z.; formal analysis, Y.D.; investigation, C.L.; resources, W.Z.; data curation, H.X.; writing—original draft preparation, Y.D.; writing—review and editing, W.Z.; visualization, C.L.; supervision, H.X.; project administration, J.T.; funding acquisition, J.T. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially supported by NSFC under grants No. 62302513.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The dataset can be obtained from the Github repository: https://github.com/TimDettmers/ConvE (accessed on 15 January 2024) and https://github.com/xwhan/DeepPath (accessed on 15 January 2024).

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Zhang, Y.; Dai, H.; Kozareva, Z.; Smola, A.J.; Song, L. Variational Reasoning for Question Answering With Knowledge Graph. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence (AAAI 2018), New Orleans, LA, USA, 2–7 February 2018; pp. 6069–6076. [Google Scholar]
  2. Verlinden, S.; Zaporojets, K.; Deleu, J.; Demeester, T.; Develder, C. Injecting Knowledge Base Information into End-to-End Joint Entity and Relation Extraction and Coreference Resolution. In Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021; Association for Computational Linguistics: Kerrville, TX, USA, 2021; pp. 1952–1957. [Google Scholar]
  3. Wang, H.; Zhao, M.; Xie, X.; Li, W.; Guo, M. Knowledge Graph Convolutional Networks for Recommender Systems. In Proceedings of the WWW 2019, San Francisco, CA, USA, 13–17 May 2019; pp. 3307–3313. [Google Scholar]
  4. Zhao, X.; Zeng, W.; Tang, J. Entity Alignment—Concepts, Recent Advances and Novel Approaches; Springer: Singapore, 2023. [Google Scholar] [CrossRef]
  5. Zeng, W.; Zhao, X.; Li, X.; Tang, J.; Wang, W. On entity alignment at scale. VLDB J. 2022, 31, 1009–1033. [Google Scholar] [CrossRef]
  6. Bollacker, K.D.; Evans, C.; Paritosh, P.K.; Sturge, T.; Taylor, J. Freebase: A collaboratively created graph database for structuring human knowledge. In Proceedings of the SIGMOD Conference 2008, Vancouver, BC, Canada, 10–12 June 2008; pp. 1247–1250. [Google Scholar]
  7. Vrandecic, D. Wikidata: A new platform for collaborative data collection. In Proceedings of the WWW 2012, Lyon, France, 16–20 April 2012; pp. 1063–1064. [Google Scholar]
  8. Bordes, A.; Usunier, N.; García-Durán, A.; Weston, J.; Yakhnenko, O. Translating Embeddings for Modeling Multi-relational Data. In Proceedings of the NIPS 2013, Lake Tahoe, NV, USA, 5–10 December 2013; pp. 2787–2795. [Google Scholar]
  9. Trouillon, T.; Welbl, J.; Riedel, S.; Gaussier, É.; Bouchard, G. Complex Embeddings for Simple Link Prediction. In Proceedings of the 33rd International Conference on International Conference on Machine Learning (ICML 2016), New York, NY, USA, 19–24 June 2016; Volume 48, pp. 2071–2080. [Google Scholar]
  10. Schlichtkrull, M.S.; Kipf, T.N.; Bloem, P.; van den Berg, R.; Titov, I.; Welling, M. Modeling Relational Data with Graph Convolutional Networks. In The Semantic Web, Proceedings of the 15th International Conference, ESWC 2018, Heraklion, Crete, Greece, 3–7 June 2018; Proceedings 15; Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2018; Volume 10843, pp. 593–607. [Google Scholar]
  11. Vashishth, S.; Sanyal, S.; Nitin, V.; Talukdar, P.P. Composition-based Multi-Relational Graph Convolutional Networks. In Proceedings of the ICLR 2020, Addis Ababa, Ethiopia, 26–30 April 2020. [Google Scholar]
  12. Liu, J.; Fan, C.; Zhou, F.; Xu, H. Complete feature learning and consistent relation modeling for few-shot knowledge graph completion. Expert Syst. Appl. 2024, 238, 121725. [Google Scholar] [CrossRef]
  13. Lin, Q.; Liu, J.; Xu, F.; Pan, Y.; Zhu, Y.; Zhang, L.; Zhao, T. Incorporating Context Graph with Logical Reasoning for Inductive Relation Prediction. In Proceedings of the SIGIR 2022, Madrid, Spain, 11–15 July 2022; pp. 893–903. [Google Scholar]
  14. Yang, F.; Yang, Z.; Cohen, W.W. Differentiable Learning of Logical Rules for Knowledge Base Reasoning. In Proceedings of the NIPS 2017, Long Beach, CA, USA, 4–9 December 2017; pp. 2319–2328. [Google Scholar]
  15. Teru, K.K.; Denis, E.G.; Hamilton, W.L. Inductive Relation Prediction by Subgraph Reasoning. In Proceedings of the ICML 2020, Virtual, 13–18 July 2020; Volume 119, pp. 9448–9457. [Google Scholar]
  16. Chen, J.; He, H.; Wu, F.; Wang, J. Topology-Aware Correlations Between Relations for Inductive Link Prediction in Knowledge Graphs. In Proceedings of the AAAI 2021, Virtual, 2–9 February 2021; pp. 6271–6278. [Google Scholar]
  17. Mai, S.; Zheng, S.; Yang, Y.; Hu, H. Communicative Message Passing for Inductive Relation Reasoning. In Proceedings of the AAAI 2021, Virtual, 2–9 February 2021; pp. 4294–4302. [Google Scholar]
  18. Galkin, M.; Denis, E.G.; Wu, J.; Hamilton, W.L. NodePiece: Compositional and Parameter-Efficient Representations of Large Knowledge Graphs. In Proceedings of the ICLR 2022, Virtual, 25–29 April 2022. [Google Scholar]
  19. Baek, J.; Lee, D.B.; Hwang, S.J. Learning to Extrapolate Knowledge: Transductive Few-shot Out-of-Graph Link Prediction. In Proceedings of the NeurIPS 2020, Virtual, 6–12 December 2020. [Google Scholar]
  20. Zhang, Y.; Wang, W.; Chen, W.; Xu, J.; Liu, A.; Zhao, L. Meta-Learning Based Hyper-Relation Feature Modeling for Out-of-Knowledge-Base Embedding. In Proceedings of the CIKM 2021, Gold Coast, QLD, Australia, 1–5 November 2021; pp. 2637–2646. [Google Scholar]
  21. Xu, X.; Zhang, P.; He, Y.; Chao, C.; Yan, C. Subgraph Neighboring Relations Infomax for Inductive Link Prediction on Knowledge Graphs. In Proceedings of the IJCAI 2022, Vienna, Austria, 23–29 July 2022; pp. 2341–2347. [Google Scholar]
  22. Zeng, W.; Zhao, X.; Tang, J.; Lin, X. Collective Entity Alignment via Adaptive Features. In Proceedings of the 36th IEEE International Conference on Data Engineering, ICDE 2020, Dallas, TX, USA, 20–24 April 2020; pp. 1870–1873. [Google Scholar]
  23. Zeng, W.; Zhao, X.; Tang, J.; Lin, X.; Groth, P. Reinforcement Learning-based Collective Entity Alignment with Adaptive Features. ACM Trans. Inf. Syst. 2021, 39, 26:1–26:31. [Google Scholar] [CrossRef]
  24. Wang, J.; Lin, X.; Huang, H.; Ke, X.; Wu, R.; You, C.; Guo, K. GLANet: Temporal knowledge graph completion based on global and local information-aware network. Appl. Intell. 2023, 53, 19285–19301. [Google Scholar] [CrossRef]
  25. Meng, X.; Bai, L.; Hu, J.; Zhu, L. Multi-hop path reasoning over sparse temporal knowledge graphs based on path completion and reward shaping. Inf. Process. Manag. 2024, 61, 103605. [Google Scholar] [CrossRef]
  26. Wang, J.; Wang, B.; Gao, J.; Hu, S.; Hu, Y.; Yin, B. Multi-Level Interaction Based Knowledge Graph Completion. IEEE ACM Trans. Audio Speech Lang. Process. 2024, 32, 386–396. [Google Scholar] [CrossRef]
  27. Sadeghian, A.; Armandpour, M.; Ding, P.; Wang, D.Z. DRUM: End-To-End Differentiable Rule Mining on Knowledge Graphs. In Proceedings of the NeurIPS 2019, Vancouver, BC, Canada, 8–14 December 2019; pp. 15321–15331. [Google Scholar]
  28. Li, J.; Wang, Q.; Mao, Z. Inductive Relation Prediction from Relational Paths and Context with Hierarchical Transformers. In Proceedings of the ICASSP 2023, Rhodes Island, Greece, 4–10 June 2023; pp. 1–5. [Google Scholar]
  29. Geng, Y.; Chen, J.; Pan, J.Z.; Chen, M.; Jiang, S.; Zhang, W.; Chen, H. Relational Message Passing for Fully Inductive Knowledge Graph Completion. In Proceedings of the ICDE 2023, Anaheim, CA, USA, 3–7 April 2023; pp. 1221–1233. [Google Scholar]
  30. Lenat, D.B. CYC: A Large-Scale Investment in Knowledge Infrastructure. Commun. ACM 1995, 38, 32–38. [Google Scholar] [CrossRef]
  31. Suchanek, F.M.; Kasneci, G.; Weikum, G. Yago: A core of semantic knowledge. In Proceedings of the WWW 2007, Banff, AB, Canada, 8–12 May 2007; pp. 697–706. [Google Scholar]
  32. Song, Y.; Zheng, S.; Niu, Z.; Fu, Z.; Lu, Y.; Yang, Y. Communicative Representation Learning on Attributed Molecular Graphs. In Proceedings of the IJCAI 2020, Yokohama, Japan, 7–15 January 2021; pp. 2831–2838. [Google Scholar]
  33. Dettmers, T.; Minervini, P.; Stenetorp, P.; Riedel, S. Convolutional 2D Knowledge Graph Embeddings. In Proceedings of the AAAI 2018, New Orleans, LO, USA, 2–7 February 2018; pp. 1811–1818. [Google Scholar]
  34. Toutanova, K.; Chen, D.; Pantel, P.; Poon, H.; Choudhury, P.; Gamon, M. Representing Text for Joint Embedding of Text and Knowledge Bases. In Proceedings of the EMNLP 2015, Lisbon, Portugal, 17–21 September 2015; pp. 1499–1509. [Google Scholar]
  35. Xiong, W.; Hoang, T.; Wang, W.Y. DeepPath: A Reinforcement Learning Method for Knowledge Graph Reasoning. In Proceedings of the EMNLP 2017, Copenhagen, Denmark, 7–11 September 2017; pp. 564–573. [Google Scholar]
Figure 1. An explanatory case in inductive relation prediction, which learned from a (a) training graph, and generalizes to be (b) without any shared entities for inference. A red dashed line denotes the relation to be predicted.
Figure 1. An explanatory case in inductive relation prediction, which learned from a (a) training graph, and generalizes to be (b) without any shared entities for inference. A red dashed line denotes the relation to be predicted.
Applsci 14 02044 g001
Figure 2. The overall framework of CNIA, which consists of the following parts: (1) extraction of the subgraphs with complete neighboring relations and initialization of the node features; (2) the feeding of subgraphs into the SNN with dual attention mechanisms to learn representations and extract neighboring relational path features; (3) commonsense-based re-ranking; (4) maximizing the MI between subgraph to graph to model the neighboring relations in a global way.
Figure 2. The overall framework of CNIA, which consists of the following parts: (1) extraction of the subgraphs with complete neighboring relations and initialization of the node features; (2) the feeding of subgraphs into the SNN with dual attention mechanisms to learn representations and extract neighboring relational path features; (3) commonsense-based re-ranking; (4) maximizing the MI between subgraph to graph to model the neighboring relations in a global way.
Applsci 14 02044 g002
Figure 3. An example of the neighbor relational paths, which are denoted by orange arrows, connecting the head to tail. And the relation context is denoted by blue edges.
Figure 3. An example of the neighbor relational paths, which are denoted by orange arrows, connecting the head to tail. And the relation context is denoted by blue edges.
Applsci 14 02044 g003
Figure 4. Discussion of the λ values in the WN18RR dataset.
Figure 4. Discussion of the λ values in the WN18RR dataset.
Applsci 14 02044 g004
Figure 5. Learning rate discussion on WN18RR.
Figure 5. Learning rate discussion on WN18RR.
Applsci 14 02044 g005
Table 1. Definitions of the notations.
Table 1. Definitions of the notations.
SymbolsDescriptions
u , r t , v Target triple
r t Target relationship to be predicted
G u , r t , v Extracted subgraph for target triple
h G The representation of subgraph G
V G Node set of subgraph G
λ The weight of MI maximization mechanism
p G The relational path representation of the subgraph
s G The representation of subgraph G
β u , r k Edge-aware attention
γ r , r t k Relation-aware attention
α u , r t , v k Joint neighbor attention
c h , r t , c t Common sense
ω The weight of commonsense in score function
Table 2. The statistics of three inductive datasets. #E and #R, respectively, denote the number of entities and relations. #TR1 represents triples, which are utilized to form KG G, while triples in #TR2 are utilized to evaluate.
Table 2. The statistics of three inductive datasets. #E and #R, respectively, denote the number of entities and relations. #TR1 represents triples, which are utilized to form KG G, while triples in #TR2 are utilized to evaluate.
VersionSplitWN18RRFB15k-237NELL-995
#R#E#TR1#TR2#R#E#TR1#TR2#R#E#TR1#TR2
v1Train927465410630183200042454891410,9154687414
Test992216181881461500199320514225833100
v2Train10695415,26218382033000973911668825648219922
Test1029234011441176200041454787949374586476
v3Train1112,07825,9013097218400017,9862194142464716,3931851
Test11508463276051873000740686512249218048809
v4Train938617940934222500027,20333527720927546876
Test9720812,3341429204350011,71414246132947073731
Table 3. The AUC-PR metric values (%) of inductive relation predictions on twelve dataset versions.
Table 3. The AUC-PR metric values (%) of inductive relation predictions on twelve dataset versions.
ModelWN18RRFB15k-237NELL-995
v1v2v3v4Avg.v1v2v3v4Avg.v1v2v3v4Avg.
GraIL94.3294.1885.8092.7291.7584.6990.5791.6894.4690.3586.0592.6293.3487.5089.87
TACT95.4397.5487.6596.0494.1683.1593.0192.1094.2590.6281.0693.1296.0785.7589.00
CoMPILE98.2399.5693.6099.8097.7985.5091.6893.1294.9091.3080.1695.8896.0885.4889.04
SNRI99.1099.9294.9099.6198.3886.6991.7791.2293.3790.76
ConGLR99.5899.6793.7899.8898.2285.6892.3293.9195.0591.7486.4895.2296.1688.4691.58
RMPI95.0095.9688.5395.7893.8285.2592.1992.0992.8090.5881.1293.4695.3591.7790.43
CNIA (ours)99.8999.9194.9599.9098.6686.9992.7593.4294.8091.9995.5095.4396.2588.1393.83
Table 4. The Hits@10 metric values (%) of inductive relation predictions on twelve dataset versions.
Table 4. The Hits@10 metric values (%) of inductive relation predictions on twelve dataset versions.
ModelWN18RRFB15k-237NELL-995
v1v2v3v4Avg.v1v2v3v4Avg.v1v2v3v4Avg.
GraIL82.4578.6858.4373.4173.2464.1581.8082.8389.2979.5159.5093.2591.4173.1979.33
TACT84.0481.6367.9776.5677.5565.7683.5685.2088.6980.8079.8088.9194.0273.7884.12
CoMPILE83.6079.8260.6975.4974.9067.6482.9884.6787.4480.6858.3893.8792.7775.1980.05
SNRI87.2383.1067.3183.3280.2471.7986.5089.5989.3984.32
ConGLR85.6492.9370.7492.9085.5568.2985.9888.6189.3182.9381.0794.9294.3681.6187.99
RMPI82.4578.6858.6873.4173.3165.3781.8081.1087.2578.8859.5092.2393.5787.6283.23
CNIA (ours)89.3685.9472.0683.0382.6074.2086.5189.4389.3984.8885.2294.1194.5582.0388.98
Table 5. The Hits@10 metric values (%) of the ablation experiment.
Table 5. The Hits@10 metric values (%) of the ablation experiment.
AblationWN18RR
v1v2v3v4
CNIA89.3685.9472.0683.03
CNIA w/o Att85.6484.6966.6982.40
CNIA w/o NRF86.4585.0671.4072.98
CNIA w/o NRP84.1685.1071.8174.07
CNIA w/o CSR86.4483.9070.0082.15
CNIA w/o CL85.1280.9564.4973.79
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Duan, Y.; Tang, J.; Xu, H.; Liu, C.; Zeng, W. Commonsense-Guided Inductive Relation Prediction with Dual Attention Mechanism. Appl. Sci. 2024, 14, 2044. https://doi.org/10.3390/app14052044

AMA Style

Duan Y, Tang J, Xu H, Liu C, Zeng W. Commonsense-Guided Inductive Relation Prediction with Dual Attention Mechanism. Applied Sciences. 2024; 14(5):2044. https://doi.org/10.3390/app14052044

Chicago/Turabian Style

Duan, Yuxiao, Jiuyang Tang, Hao Xu, Changsen Liu, and Weixin Zeng. 2024. "Commonsense-Guided Inductive Relation Prediction with Dual Attention Mechanism" Applied Sciences 14, no. 5: 2044. https://doi.org/10.3390/app14052044

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop