Next Article in Journal
Adaptive Outdoor Cleaning Robot with Real-Time Terrain Perception and Fuzzy Control
Next Article in Special Issue
Lightweight Attention-Based CNN Architecture for CSI Feedback of RIS-Assisted MISO Systems
Previous Article in Journal
Equivalence Transformation for Neutral Differential Equations: Oscillation of Solutions
Previous Article in Special Issue
Wi-FiAG: Fine-Grained Abnormal Gait Recognition via CNN-BiGRU with Attention Mechanism from Wi-Fi CSI
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Large Language Models for Knowledge Graph Embedding: A Survey

1
School of Software, Shandong University, Jinan 250101, China
2
Department of Computer Science, Metropolitan College, Boston University, Boston, MA 02215, USA
3
Fujian Key Laboratory of Sensing and Computing for Smart City, School of Informatics, Xiamen University, Xiamen 361005, China
*
Authors to whom correspondence should be addressed.
Mathematics 2025, 13(14), 2244; https://doi.org/10.3390/math13142244
Submission received: 16 May 2025 / Revised: 29 June 2025 / Accepted: 1 July 2025 / Published: 10 July 2025
(This article belongs to the Special Issue Data-Driven Decentralized Learning for Future Communication Networks)

Abstract

Large language models (LLMs) have attracted a lot of attention in various fields due to their superior performance, aiming to train hundreds of millions or more parameters on large amounts of text data to understand and generate natural language. As the superior performance of LLMs becomes apparent, they are increasingly being applied to knowledge graph embedding (KGE)-related tasks to improve the processing results. Traditional KGE representation learning methods map entities and relations into a low-dimensional vector space, enabling the triples in the knowledge graph to satisfy a specific scoring function in the vector space. However, based on the powerful language understanding and semantic modeling capabilities of LLMs, which have recently been invoked to varying degrees in different types of KGE-related scenarios such as multi-modal KGE and open KGE according to their task characteristics, researchers are increasingly exploring how to integrate LLMs to enhance knowledge representation, improve generalization to unseen entities or relations, and support reasoning beyond static graph structures. In this paper, we investigate a wide range of approaches for performing LLMs-related tasks in different types of KGE scenarios. To better compare the various approaches, we summarize each KGE scenario in a classification. In the article we also discuss the applications in which the methods are mainly used and suggest several forward-looking directions for the development of this new research area.

1. Introduction

Knowledge graphs (KGs) are a technology that organizes and represents knowledge in the form of a graph structure, describing complex real-world associations through entities, attributes, and relations. With the development of society and the advancement of science and technology, KGs have been more and more widely used in applications containing rich semantic information in different fields. Compared to traditional data analysis methods, KGs can more intuitively model heterogeneous relations and enable interpretable reasoning across various application domains [1]. At present, KGs have been widely used in various fields such as social networks [2,3], bioinformatics networks [4,5], and traffic flow prediction [6,7], proved to have superior performance. Traditional approaches to handling the aforementioned tasks typically rely on learning knowledge graph embedding (KGE) representations as a foundational step. KG organize factual knowledge as structured semantic networks, while KGE provide differentiable vector-space representations that preserve KG’s topological and relational properties for downstream analytic tasks. To facilitate the reader’s understanding of the relationship between KG and KGE, we provide an example here. Suppose that we have a KG that contains entities “Elon Musk” and “Tesla Inc.”, as well as the relationship “founded”. The task of the KGE is to map these entities and relations into a low-dimensional vector space. For example, the embedding vector for “Elon Musk” might be ‘[0.1, 0.2, 0.3]’, for “founded” it might be ‘[0.4, 0.5, 0.6]’, and for “Tesla Inc.” it might be ‘[0.7, 0.8, 0.9]’. Through this embedding, the triples in the KG can satisfy certain semantic relations in the vector space, such as ‘embedding (Elon Musk) + embedding (founded) ≈ embedding (Tesla Inc.)’. This embedding method not only preserves the structural information of the KG but also supports complex reasoning and analytical tasks. In the example, the dimensions and initial values of the entity mapping vectors are generated based on predefined rules instead of being random, and such low-dimensional embeddings can effectively preserve the structural information of the knowledge graph to support complex reasoning and analytical tasks. Subsequently, relevant tasks are carried out on the basis of these precise KGE representations. However, deep learning models in traditional approaches often suffer from performance limitations due to their training scale. Despite the need for large amounts of text data for training, LLMs are increasingly being applied to KGE-related tasks due to their multiple advantages. Pre-trained on massive text data, LLMs can capture rich linguistic patterns and semantic knowledge, which can be effectively transferred to KGE tasks to provide a strong foundation for learning relationships and structures within knowledge graphs.
A large language model (LLM) is a kind of deep learning model trained on massive text data that can understand and generate human language, and it can perform various natural language processing tasks (such as translation, text generation, etc.) through contextual learning. LLMs are increasingly being applied to KGE-related tasks to enhance their processing because of their superior performance. LLMs can be built upon datasets that are several times larger than those of traditional models. When combined with techniques such as fine-tuning and the use of calling cue words, they can achieve even more remarkable performance in KGE tasks compared to traditional predictive models [8,9,10,11,12]. The aim of this paper is to provide a comprehensive overview and structural categorization of the various task approaches that have been applied to LLMs in different KGE scenarios.
Unlike previous surveys that may have provided fragmented or narrow coverage, this survey aims to offer a truly comprehensive overview of the work related to various KGE scenarios within the context of LLMs. By approaching the topic from the perspective of different methods of calling LLMs, it comprehensively encompasses all aspects of KGE application scenarios and use cases, presenting a more in-depth and holistic analysis that fills the gaps left by previous works. The previous reviews [13] try to classify LLMs and KGE from the perspective of enhancement methods, or some reviews [14] classify the different stages of KG enhancement and pre-trained language model enhancement based on LLMs. With the continuous progress of research, various emerging application scenarios for KGE are emerging. However, there is still a lack of LLM classification overview of KGE representation learning in various application scenarios. There have also been a number of surveys about KGE or LLM before. Some articles have been investigated only for KGE aspects. For example, the review written by Liu et al. [15] discusses in detail the current application status, technical methods, practical applications, challenges, and future prospects of KGEs in industrial control system security situational awareness and decision-making. A review by Zhu et al. [16] summarized the KG entity alignment method based on representation learning, proposed a new framework, and compared the performance of different models on diverse datasets in detail. Some articles have been investigated only for some applications related to LLMs. For example, a review by Huang et al. [17] explores the security and credibility of LLMs in industrial applications, analyzes vulnerabilities and limitations, explores integrated and extended verification and validation techniques, and contributes to the full lifecycle security of LLMs and its applications. Kumar, P [18] comprehensively explores the applications of LLMs in language modeling, word embedding, and deep learning, analyzes their diverse applications in multiple fields and the limitations of existing methods, and proposes future research directions and potential progress. The aforementioned studies primarily focus on isolated aspects of KGE applications. In contrast, our work presents a comprehensive investigation into the novel integration of large language models (LLMs) with knowledge graph embeddings (KGEs).
While Liu et al. [15] provided a detailed analysis of KGE applications in industrial control systems, covering technical approaches and practical challenges, their work was limited to a single domain without exploring the synergistic potential with LLMs, resulting in insufficient generalizability. Zhu et al. [16] systematically summarized representation learning-based KG entity alignment methods, yet failed to consider LLMs’ dynamic reasoning capabilities in entity alignment. Huang et al. [17] conducted an in-depth examination of security risks in LLM industrial applications and proposed lifecycle verification techniques with strong practical guidance, but they completely overlooked the role of KGE in enhancing LLM trustworthiness (e.g., knowledge verification). Kumar [18] comprehensively reviewed LLM potential in language modeling and cross-domain applications, but lacked systematic analysis of KGE-LLM integration.
Integrating LLMs into KGE tasks brings significant advantages. The pre-trained knowledge of LLMs provides a strong semantic foundation for KGE models, reducing the number of iterations required for training from scratch and thereby significantly shortening training time and accelerating model convergence. Meanwhile, the incorporation of LLMs lowers the dependence on large-scale labeled data, allowing KGE models to achieve good performance with less labeled data while enhancing data diversity and improving model generalization through their generative capabilities. In terms of hyperparameter tuning, the pre-trained structure and parameters of LLMs offer a stable reference, reducing the workload and difficulty of adjustment and improving tuning efficiency. Additionally, the efficient computing power and optimized pre-trained structure of LLMs lower energy consumption during training and inference enhance the model’s ability to handle large-scale knowledge graphs. Overall, these advantages demonstrate great application potential in knowledge graph-related tasks and bring new breakthroughs to the field of knowledge graph embedding.
Through in-depth analysis, we have identified two common limitations in current survey works, outlined as follows: (1) insufficient research on the integration of KGE and LLMs, and (2) narrow research scopes or application scenarios. In response, our study makes the following two key contributions:
  • We propose a unified methodology for KGE-LLM synergy, systematically analyzing their complementary enhancement mechanisms across diverse application scenarios (e.g., dynamic knowledge graphs and multimodal reasoning), thereby overcoming the single-perspective limitations of existing research.
  • We introduce an innovative classification framework for LLM-KGE integration technologies across broader research domains and application scenarios, addressing the gaps in systematicity and scalability in current studies.
Building upon the research groundwork and surveys presented in the aforementioned related papers, we conduct a more comprehensive literature review. Our primary focus lies on the current KGE application scenarios. We categorize and discuss these diverse KGE application scenarios based on the degree of LLM invocations. Through such in-depth descriptions, we achieve a comprehensive exploration of this domain, uncovering nuances and relations that have not been fully examined before. This paper is organized as follows: In Section 2, we introduce some basic concepts related to this paper. In Section 3, we discuss the application of LLMs in different KGE scenarios. In Section 4, we propose some possible future applications and research directions. In Section 5, we introduce some famous datasets and code resources; finally, we conclude the paper in Section 6.

2. Preliminaries

To ensure comprehensibility for the reader, we introduce relevant basic concepts in this section. We summarize all acronyms used throughout the manuscript in Table 1.

2.1. Large Language Models (LLMs)

LLMs can be defined as a type of artificial intelligence technology based on deep learning, trained on massive text corpora to generate and understand natural language. Their basic process involves extracting language patterns from large-scale data, modeling them with neural networks, and generating high-quality language outputs based on user inputs. Their applications span various fields, including text generation [19], machine translation [20], question answering [21], sentiment analysis [22], and content summarization [23]. Compared to traditional methods, LLMs offer superior contextual understanding, higher output quality, and enhanced transfer learning capabilities.

2.2. KGE-Related Tasks

In this section, we introduce several tasks related to knowledge graph embedding.

Relationship Between KG and KGE

With the development of society and the progress of science and technology, graph data have been increasingly used in various fields, such as social networks [24,25], academic networks [26], biological networks [27], etc. Graph data in various domains often consist of multiple types of edges and nodes, which contain rich semantic information, and they put high demands on accurate graph data information representation in various types of downstream applications in different domains. KGs are structured representations of knowledge that model real-world entities, their attributes, and interrelationships as nodes and edges in a semantic network, enabling formalized knowledge organization and storage. Building upon this structured foundation, KGEs generate low-dimensional, continuous vector representations by learning from KGs, effectively preserving their inherent relational patterns while transforming discrete graph structures into numerical forms suitable for machine learning and statistical inference. KGEs transform entities and relations in KGs into embedded representations mapped to a low-dimensional continuous vector space. KGEs can adequately capture the complex graph data structure and rich semantic information of various domains, facilitating the management of the huge amount of graph data in different domains, use, and maintenance. At present, in various fields such as medicine [28,29], news media [30], etc. KGEs have been widely used in related applications, and they have become one of the core technologies of today’s scientific and technological development. Currently, several widely used benchmark datasets [31,32,33] contain millions of entities and relations, including information about locations, music, and more. These datasets are commonly used for evaluation tasks across various domains. In some specific assessment tasks, specific datasets are also used, such as the COMBO dataset used in the assessment task of the open KG [34].

2.3. Classic Knowledge Graph (CKG)

In the remainder of this section we discuss several types of classic quest scenarios.

2.3.1. Link Prediction

Link prediction is a core task in KG completion, aimed at predicting missing entity relation pairs (i.e., missing edges) in a KG [35]. The goal is to predict potential new entity relation pairs based on existing entities and relations.

2.3.2. Entity Alignment

Entity alignment is a key task in KG fusion. Its goal is to find nodes (entities) representing the same real-world entities in multiple KGs and align them. For example, the “Barack Obama” entity in one KG should correspond to the ”Obama president” entity in another KG. Entity alignment is crucial for the integration of KGs, data fusion, and cross-domain graph inference.

2.3.3. KG Canonicalization

KG canonicalization is an important task in knowledge storage, sharing, retrieval, and application. Its goal is to map multiple entities in KGs to a standardized embedding representation, eliminating redundancy and ambiguity, thereby improving the quality and consistency of KGEs, for example, mapping different embedding representations such as “New York City”, “NYC”, and “Big Apple” to a unified embedding representation. KG canonicalization typically involves two main steps. The first step involves learning accurate embedding representations. In the second step, entities are clustered based on the similarity of their learned embeddings, and a unified representation is assigned to each cluster.
A concrete example is provided in Figure 1 for reference as follows: “Musk”, “Elon Musk”, and “Chairman Musk” all represent the same entity, while “founded” and “created” convey identical relationships. In this task, phrases representing the same meaning can be clustered together and assigned a unified identifier to prevent redundancy and ambiguity.

3. Techniques of KGE with LLMs

We depict the differences between the above categories of KGs and traditional KGs in Figure 2. The characteristics of KGE methods vary across different KG scenarios, reflecting their unique structural and semantic requirements. Current traditional KGE methods face limitations in handling complex relations, long-tail entities, and dynamic knowledge updates, and they struggle to effectively capture semantic information and contextual associations. LLMs can compensate for the shortcomings of traditional KGE in semantic expression and contextual reasoning through their powerful semantic understanding and generation capabilities, thereby enhancing the accuracy of knowledge representation and reasoning. With the introduction of LLMs, novel approaches combining pre-trained language models have emerged in the KGE field, demonstrating significant advantages and development potential in knowledge completion, relation prediction, and cross-domain applications. We classify the utilization of LLMs based on various KGE application scenarios and elaborate on this in the following section. In Table 2, we present a comparative analysis of the six distinct types of KGs discussed in our study, examining their key characteristics, application scenarios, as well as their respective strengths and limitations.

3.1. Classic Knowledge Graph (CKG)

A CKG is a structured semantic knowledge base based on graphs, which stores graph data in the form of a triplet structure. Currently, CKGSs have been widely used in various task categories in various fields. As research continues to deepen and new application requirements emerge, several KG scenarios distinct from CKGs have been proposed. These scenarios will be introduced in detail in the following sections. To distinguish it from knowledge graphs in other scenarios, we refer to this type as the classic knowledge graph (CKG) throughout the remainder of this paper.
Based on the extent to which large language models (LLMs) are utilized, KGE methods within the CKG scenario can be broadly categorized into four types as follows: (1) methods that use LLMs’ prompts; (2) methods for fine-tuning LLMs; (3) methods of pre-trained LLMs; and (4) methods to use LLMs as agents. This classification reflects the varying degrees of LLM integration in these approaches. A comparative summary of these four categories—highlighting their workflows, strengths, and limitations—is presented in Table 3.

3.1.1. Methods That Use LLMs’ Prompts

Initially, the structural information and auxiliary details of triplets are extracted and then transformed into formatted text descriptions as prompt inputs for large language models (LLMs). By leveraging the text-generation and knowledge-reasoning capabilities of LLMs, tasks based on Knowledge Graph Embedding (KGE) methods can be completed. For instance, KICGPT [36] adopts a “knowledge cues” contextual learning strategy, encoding structured knowledge into representations to guide LLMs. FireRobBrain [37] integrates a prompt module to assist LLMs in generating suggestions for bot responses. The combination of LLMs and contextual knowledge graph (CKG) embeddings—facilitated by a well-designed cue module—significantly enhances answer quality. PEG [38] utilizes prompt-based ensemble KGE, augmenting existing methods through knowledge acquisition, aggregation, and injection, refining evidence via semantic filtering, and aggregating it into global KGE through implicit or explicit approaches. KC-GenRe [39] performs knowledge-enhanced constrained inference for CKG completion, supporting contextual prompts and controlled generation to achieve effective rankings. CP-KGC [40] adapts prompts to different datasets to enhance semantic richness, using a contextual constraint strategy to effectively identify multi-aggregated entities. KGCD [41] employs an LLM-driven distillation framework with mixed prompts, which enhances the completion of low-resource knowledge graphs through LLM prompt utilization.
This type of method can effectively utilize the powerful text generation and inference capabilities of LLMs, integrating structured triplet knowledge embeddings from CKGs in a more natural way, thereby providing assistance for complex tasks such as knowledge completion, and potentially uncovering hidden semantic relations between triplets. However, there may be information loss or semantic bias during the process of converting triplets into textual descriptions. Because the triplet structure of CKGs is precise, the accuracy of the original structural information may decrease after conversion to text due to factors such as linguistic ambiguity.

3.1.2. Methods for Fine-Tuning LLMs

First, CKG triplet data are collected and transformed into a format highly compatible with model training requirements. The preprocessed triplet text data are then fed into LLMs. Suitable loss functions are defined, and model parameters are strictly fine-tuned according to specific task objectives to optimize the model for accurate and efficient task execution. To improve performance while reducing the amount of instruction data, DIFT [42] employs a truncated sampling method to select useful facts for fine-tuning and injects KG embeddings into the LLM. RecIn-Dial [43] fine-tunes large-scale pre-trained LLMs to generate fluent and diverse responses, introducing knowledge-aware biases learned from entity-oriented CKGs to enhance recommendation performance. KG-LLM [44] converts structured CKG data into natural language, using these prompts to fine-tune LLMs for improved embedding-based multi-hop link prediction. RPLLM [45] leverages CKG node names to fine-tune LLMs for relational prediction, enabling the model to operate effectively in low-resource settings using only node names. KG-Adapter [46] is a parameter-efficient CKG ensemble method based on prompt-tuning, introducing a novel adapter structure for decoder LLMs that encodes CKGs from node and relation-centric perspectives for joint inference with LLMs. LPKG [47] augments LLMs with CKG-derived data, fine-tuning them to adapt to complex KG scenarios. Filippo Bianchini et al. [48] propose a Retrieval Augmented Generation (RAG) pipeline for complex legal linguistic tasks, integrating two Knowledge Graphs (KGs) and leveraging the iterative fine-tuning of Large Language Models (LLMs). This approach addresses the fallacies from sub-symbolic processing in conventional LLMs by fine-tuning LLMs with knowledge integration.
Through targeted fine-tuning, the model learns the semantic relations and knowledge embedded in CKG triplet structures, improving its performance in CKG tasks. This approach effectively utilizes CKG data to enhance the model’s expertise in knowledge graph domains. It is important to note that fine-tuning requires substantial high-quality annotated data. Collecting and organizing CKG triplet data into suitable formats is time-consuming and labor-intensive. Inadequate data quality or quantity may lead to model overfitting, resulting in poor generalization on unseen real-world data.

3.1.3. Methods of Pre-Trained LLMs

CKG triplets and relevant background information are first transformed into a text format suitable for subsequent processing, which is then used for pre-training. For the specially designed pre-training tasks, model parameters are updated by calculating the loss between predicted results and true labels, enabling the model to effectively learn language patterns and semantic relation embeddings within the CKG. Once knowledge graph triples and background information are converted to text, parameters are updated through tasks such as masked language modeling (MLM) and next sentence prediction (NSP). Take MLM as an example; by predicting masked entity or relation tokens in sentences, the model learns semantic associations between KG entities (e.g., drug-indication relationships). The NSP task helps the model understand the logical order between triples. LiNgGuGKD [49] adopts label-oriented instruction tuning for pre-trained LLM teachers, introducing a hierarchical adaptive contrastive distillation strategy to align node features between teacher and student networks in the latent space, thus effectively transferring the understanding of semantics and complex relations. GLAME [50] utilizes public pre-trained LLMs with a KG augmentation module to discover edited relevant knowledge and obtain its internal embedding in LLMs. Salnikov et al. [51] explore text-to-text pre-trained LLMs, enriching them with CKG information to answer factual questions. PDKGC [52] employs pre-trained language models for KGC by freezing the PLM and training two prompts—a hard task prompt and a disentangled structure prompt—to enable the PLM to integrate text and graph structure knowledge, falling under the category of pre-trained LLM methods.
During pre-training, LLMs capture semantic relations and language patterns in CKG, enhancing their understanding and embedding representation of CKG while enriching knowledge reserves. However, converting CKG to text may introduce noise or information loss, affecting pre-training effectiveness and making acquired knowledge less accurate.

3.1.4. Methods to Use LLMs as Agents

The use of LLMs as agents enables the interactive exploration of relevant entities and relations on CKGs, along with inference based on retrieved knowledge. However, this approach remains underutilized and warrants further exploration in future research. ToG [53] employs an LLM agent to iteratively perform beam search on the KG, identifying the most promising inference paths and returning the most likely inference results. ODA [10] integrates LLMs with KGs as an observation-driven agent, leveraging a cyclical paradigm of observation, action, and reflection to enhance reasoning. This framework falls under the category of methods that utilize LLMs as agents for knowledge graph interaction.
When LLMs play the role of an agent, they can take advantage of the large amount of knowledge and reasoning patterns obtained in the pre-training process, and combine these with the structured information in CKG to perform more complex knowledge reasoning.

3.2. Temporal Knowledge Graph (TKG)

TKG is a kind of dynamic KG that contains facts that change over time. At present, LLMs mainly use the method based on prompts in the field of TKG for KGE tasks, accompanied by local use based on fine-tuning and other methods. In TKG, the processing of temporal information is crucial for prompt construction. For example, to query the efficacy changes of a drug within a specific time period, a prompt like this can be constructed: “Considering the time interval from the start time to the end time, analyze the efficacy data of the drug name in each quarter, including changes in indicators such as cure rate and remission rate. Please provide a dynamic evolution analysis of the drug’s efficacy based on the relevant entities and relationships in the knowledge graph within this time range.” By clarifying the time interval and specific temporal granularity, the model is guided to focus on knowledge of a specific period, effectively using the temporal information in TKG for reasoning and analysis.
In the LLM method based on prompts, the data in the TKG is generally processed first, so that it can be effectively understood and processed by LLMs, especially for the processing of time information. The prompts are usually constructed according to the specific time characteristics and other effective characteristics of the time series task. In TKG’s question–answering tasks, GenTKGQA [54] guides LLMs to answer the temporal question through two phases: subgraph retrieval and answer generation. The virtual knowledge index is designed, which is helpful for the open-source LLMs to deeply understand the temporal order and structural dependence between the retrieved facts through instruction adjustment. In the temporal relation prediction task, zrLLM [55] inputs textual descriptions of KG relations into LLMs to generate relation embedding representations, and then introduces them into an embedding-based approach. The representation of LLMs authorization can capture the semantic information in the description of the relation. This makes relations with similar semantics, close in the embedding space, enabling the model to recognize zero-shot relations even when no context observes. CoH [56] enhances the prediction ability of LLMs by using higher-order history chains step by step, gradually provides high-order historical information, and fuses it with graph neural network (GNN) results to obtain more accurate prediction results. Unlike previous tasks, GenTKG [57] also uses fine-tuning, using an instruction fine-tuning strategy with few sample parameters to align with the task of LLMs, and using a retrieval enhancement generation framework to solve the challenge through a two-stage approach.
The prompt-based method does not require a large amount of training data and complex fine-tuning process in the TKG task, and can directly use the ability of the pre-trained LLMs to guide the model to generate results that meet the requirements through well-designed prompts, which can effectively reduce the cost and time of the task. However, the design and selection of prompts are highly demanding, and if the prompts are not accurate, clear or comprehensive, the model may generate inaccurate or irrelevant results, and different prompts need to be designed for different tasks, and the universality is limited. In contrast, the fine-tuning-based method can optimize the pre-trained model according to specific tasks and datasets in TKG-related tasks, so that the model can better adapt to specific fields and tasks and improve the performance and accuracy of the model on this task, which is worthy of further exploration in the future. TPNet [58] introduces a pre-trained language model (PLM) to jointly encode temporal paths with a time-aware encoder, addressing the failure to handle continuously emerging unseen entities and ignore historical dependencies in traditional TKGR methods. DSEP [59] employs a pre-trained language model (PLM) to learn the evolutionary characteristics of historical related facts, addressing the neglect of inherent semantic information in historical facts by existing TKGR methods.

3.3. Multi-Modal Knowledge Graph (MMKG)

A multi-modal knowledge graph is a kind of KG that combines multiple data modalities (e.g., text, image, video, etc.) to construct and represent facts. MMKG enhances the diversity and accuracy of knowledge representation by enriching modal data. For MMKG, integrating data from different modalities is key. Take a medical knowledge graph containing text and images as an example. To identify potential risk factors for a specific disease, a prompt can be designed as follows: “The text describes patient information including symptoms and medical history, with key details such as long-term smoking history and family genetic history listed. Meanwhile, the provided medical images show critical features like the shape, size, and location of lung shadows. Please combine the textual and image information, infer the patient’s potential disease risk factors based on entity relationships in the knowledge graph, and provide relevant justifications”. In this way, by separately presenting the core content of multi-modal data and emphasizing reasoning by integrating both, the model can fully leverage diverse data in MMKG to enhance the accuracy and comprehensiveness of task execution.
Currently, prompt-based LLM methods are primarily applied in the domain of MMKGs. Features from different modalities are extracted and encoded into prompts, which are then fed into LLMs to perform multi-modal fusion and reasoning, enabling the completion of downstream tasks. In the KG completion task, MPIKGC [60] prompts the LLMs to generate auxiliary text to improve the performance of the model. MML [61] initiates the process by meticulously constructing MMKGs with the specific aim of extracting regional embedding representations that inherently encapsulate spatial heterogeneity features. Subsequently, it deftly employs LLMs to generate diverse multi-modal features pertaining to the ecological civilization model. KPD [62] designs a prompt template with three contextual examples to understand all aspects of a web page for phishing detection. In order to construct a unified KG, Docs2KG [63] extends these nodes with multi-hop queries to retrieve relevant information, thereby enhancing the response prompts to queries.
The prompt-based LLM method can enhance the understanding and fusion ability of multi-modal information, handle complex multi-modal tasks better, and improve the application effect and value. However, the results may be inaccurate and unreliable due to the limitations of LLMs themselves, such as hallucinations and insufficient attention to modal information such as images. Suggested LLMs based on pre-training or agents can be explored in the next step.

3.4. Inductive Knowledge Graph (IKG)

Inductive knowledge graph is a type of KG that focuses on transferring knowledge patterns from a source graph to a target graph, enabling the prediction of new entities and relations in unseen scenarios. IKG includes a source KG and a target KG with specific conditions on relations and entities. The goal of embedding representation learning for inductive KGs is to capture structural patterns from the source KG and transfer them to the target KG. At present, in the direction of IKG, LLMs are almost always applied to related tasks in the form of prompts. According to the structure and task design prompts of IKG, the relevant text and prompt words can be input into the LLMs. LLMs can infer and analyze KG according to the prompts and generate the results. Xing et al. [64] construct a fusion inductive framework to effectively alleviate the illusory problem of LLMs, combined with predefined prompts, to provide powerful and intelligent decision support for practical applications. LLAMA3 [65] uses the current generation of LLMs for link prediction, and it realizes the new entity link prediction in the IKG by generating prompts, answer processing, entity candidate search, and final link prediction. Different from the previous two attempts in the inductive link prediction task, LLM-KERec [66] innovates in the IKG of the industrial recommendation system and designs prompts covering multiple aspects. ProLINK [67] employs prompt-based LLMs to generate graph-structural prompts for enhancing pre-trained GNNs, addressing the challenge of handling low-resource scenarios with scarcity in both textual and structural aspects in KG inductive reasoning. By introducing an entity extractor, it extracts unified conceptual terms from project and user information. To provide cost-effective and reliable prior knowledge, entity pairs are generated based on the popularity of the entity and specific strategies.
The prompt-based approach can enhance IKG’s inductive ability to better handle complex textual information and reasoning tasks with the help of the language understanding and generation capabilities of LLMs. At the same time, hallucinations may occur, resulting in inaccurate or unreliable reasoning results, affecting the quality and application effect of IKGs. More attempts with fine-tuning-based or pre-training-based approaches are recommendation in the future.

3.5. Open Knowledge Graph (OKG)

OKG is a kind of KG derived from open datasets, user contributions or internet information, emphasizing the openness and scalability of knowledge. OKG lowers the threshold of knowledge access through openness and sharing, while facing the problems of data quality and reliability. With the advent of the big data era, massive data sources continue to be generated. OKGs have gained attention in more and more scenarios, such as responding to cyber threat attacks [68], mental health observation [69], and understanding geographic relations [70]. Based on the different degrees of use of such methods for LLMs, we can roughly divide them into two categories as follows: (1) methods that use LLMs’ prompts; (2) methods for fine-tuning LLMs.

3.5.1. Methods That Use LLMs’ Prompts

Specific prompts are designed to transform data in OKGs into inputs meeting the requirements. This enables LLMs to generate textual embedding representations, which are then encoded or converted into vector-form embeddings to support subsequent tasks like knowledge reasoning. Arsenyan et al. [71] propose guided prompt design for LLM utilization, constructing OKGs from electronic medical record notes. The KG entities include diseases, influencing factors, treatments, and manifestations co-occurring with patients during illness. Xu et al. [72] leverage LLMs and medical expert refinement to build a heart failure KG, applying prompt engineering across three phases: schema design, information extraction, and knowledge completion. Task-specific prompt templates and two-way chat methods achieve optimal performance. The approach by AlMahri et al. [73] automates supply chain information extraction from diverse public sources, constructing KGs to capture complex interdependencies between supply chain entities. Zero-shot prompts are used for named entity recognition and relation extraction, eliminating the need for extensive domain-specific training. Datta et al. [74] introduced a zero-shot prompt-based method using OpenAI’s GPT-3.5 model to extract hyperrelational knowledge from text, where each prompt contains a natural language query designed to elicit specific responses from pre-trained language models. Paper2lkg [75] employs prompt-based generative LLMs to automate NLP tasks in local KG construction, addressing the challenge of transforming academic papers into structured KG representations for enriching existing academic KGs.
Prompts can guide LLMs to focus on specific knowledge content and reduce redundant expressions. By accurately designing prompts, LLMs can directly cut into key information and avoid the output of irrelevant knowledge, so as to effectively filter out redundant information that may exist in OKG. However, due to the ambiguity of the expression of OKG itself, prompts may exacerbate the misunderstanding. If the prompts do not accurately fit the ambiguous knowledge statement, LLMs may generate results based on incorrect understanding, amplifying the adverse effects of the ambiguous parts of OKG.

3.5.2. Methods for Fine-Tuning LLMs

The knowledge in OKG is extracted and converted into a format suitable for LLM input, such as text sequences, followed by fine-tuning for training. This fine-tuning process enables LLMs to learn the embedding representations of entities, relations, and other knowledge in OKG, which are then applied to downstream tasks like knowledge completion and question answering. UrbanKGent [70] first designs specific instructions to help LLMs understand complex semantic relations in urban data, and it then invokes geocomputing tools and iterative optimization algorithms to enhance the LLMs’ comprehension and reasoning about geospatial relations. Finally, the LLMs are further trained via fine-tuning to better accomplish various OKG tasks. LLM-TIKG [76] leverages ChatGPT’s few-shot learning capabilities for data annotation and augmentation, creating a dataset to fine-tune a smaller language model. The fine-tuned model is used to classify collected reports by subject and extract entities and relations.
Through fine-tuning, LLMs can learn the semantic and structural information in OKG, so as to express knowledge more accurately when generating text, reducing the expression redundancy caused by inaccurate understanding of knowledge. However, if there is bias or noise in the data used in the fine-tuning process, it may lead to the LLMs overfitting these inaccurate information, consequently producing more ambiguity or misunderstanding when processing the OKG.

3.6. Domain Knowledge Graph (DKG)

DKG is a kind of KG for a specific domain, which can often record specific knowledge in a specific domain and solve problems in a specific domain in a more targeted manner. At present, DKG has been studied on how to be applied to various domains such as medicine [77,78,79] and electric power [80], and it is broadly significant for the development of specific fields. Based on the different degrees of use of such methods for LLMs, we can roughly divide them into three categories as follows: (1) methods that use LLMs’ prompts; (2) methods for fine-tuning LLMs; and (3) methods of pre-trained LLMs.

3.6.1. Methods That Use LLMs’ Prompts

Prompts play a crucial role in guiding LLMs to focus on domain-specific knowledge, often requiring the design of tailored prompt templates based on the characteristics of specific domains. This approach represents the most commonly used method in DKG research. In the medical domain’s DKG, KG-Rank [77] is employed to enhance medical question answering. As the first LLM framework for medical question answering, it integrates knowledge graph embedding (KGE) with ranking technology to generate and refine answers, applying medical prompts to identify relevant medical entities for given questions. Yang et al. [81] developed Scorpius, a conditional text generation model that generates malicious paper abstracts for promoted drugs and target diseases using prompts. By poisoning knowledge graphs, it elevates false relevance between entities, demonstrating the risk of LLMs manipulating scientific knowledge discovery. In the power engineering domain, Wang et al. [80] utilize LLMs to generate a domain-specific question–answer dataset. Their approach not only incorporates KG triplet relations into question prompts to improve answer quality but also uses triplet-extended question sets as knowledge augmentation data for LLM fine-tuning. However, these methods are limited to single-domain scenarios. In contrast to single-domain DKG applications, KGPA [82] addresses cross-domain KGE scenarios by using KGs to generate original and adversarial prompts, evaluating LLM robustness through these prompts. Neha Mohan Kumar et al. [83] proposed a hybrid malware classification system that employs prompt chaining with a cybersecurity knowledge graph. This system refines initial classifications from fine-tuned LLMs, addressing the challenge of improving detection accuracy and interpretability in malware classification.
Based on the prompt, it can focus on domain knowledge, and it can guide LLMs to focus on specific fields, such as accurately obtaining medical-related content when building a medical KG. By designing different prompts, it can also adapt to a variety of KGE tasks. However, the domain adaptability of such methods is often limited, and for some specific and highly specialized fields, prompts may not be well understood and are not provided with valid answers due to insufficient LLMs training data.

3.6.2. Methods for Fine-Tuning LLMs

In scenarios where special domain data are continuously generated, fine-tuned models can better adapt to knowledge evolution, facilitate dynamic updates of DKGs, and maintain knowledge integrity and timeliness. For the KG alignment task, GLAM [84] converts KGs into alternative text representations of tagged question–answer pairs through LLM fine-tuning. Leveraging LLM generative capabilities to create datasets, it proposes an effective retrieval-enhanced generation approach. Li et al. [85] develop a fine-tuning model for KG fault diagnosis and system maintenance, with key contributions including KG construction and fine-tuning dataset development for LLM deployment in industrial fields. Additionally, in auxiliary fault diagnosis, Liu et al. [86] embed aviation assembly KGs into LLMs. The model uses graph-structured big data in DKGs to prefix LLMs, enabling online reconfiguration via DKG-based prefix tuning and avoiding substantial computational overhead. GaLM [84] introduces an LLM fine-tuning framework that transforms knowledge graphs into text-based question–answer pairs, addressing LLMs’ limited reasoning ability over domain-specific KGs and minimizing hallucination.
Through fine-tuning, the large model can better adapt to the KG tasks in a specific domain, and it can use the annotated data in the domain to further learn and optimize, so as to improve the accuracy, recall, and other performance indicators of the tasks in the field. However, there may be a risk of overfitting, if the amount of domain data used in fine-tuning is insufficient or over-trained, the large model may overfit these limited data, resulting in poor performance in unseen domain data or real-world application scenarios, reduced generalization ability, and inability to effectively process emerging situations or data.

3.6.3. Methods for Pre-Training LLMs

The integration of DKG’s structured knowledge in the pre-training process enables LLMs to directly learn KG-contained knowledge, enhancing their understanding and mastery of domain-specific knowledge while providing a robust knowledge base for DKG tasks. In the automotive electrical systems domain, Wawrzik et al. [87] utilize a Wikipedia-derived electronic dataset for pre-training, optimizing it with multiple prompting methods and features to enable timely access to structured information. In food testing, FoodGPT [88] introduces an incremental pre-training step to inject knowledge embeddings into LLMs, proposing a method to handle structured knowledge and scanned documents. To address machine hallucination, a DKG is constructed as an external knowledge base to support LLM retrieval. FuseLinker [89] employs pre-trained LLMs to derive text embeddings for biomedical knowledge graphs, tackling the challenge of fully leveraging graph textual information to enhance link prediction performance.
The pre-trained LLM method can improve the efficiency and quality of knowledge extraction in DKG tasks, and it can directly extract knowledge such as entities and relations from specific domain texts with strong language understanding and text generation capabilities, achieving specific specialized extraction in specific domains. However, there may be knowledge illusions and errors, which can generate knowledge that does not match the facts or does not exist, resulting in the existence of false information in DKGs.

3.7. Summary of Method Evaluation

3.7.1. Effectiveness

The application of Large Language Models (LLMs) across various types of Knowledge Graphs (KGs) has shown varying degrees of effectiveness. In Classic Knowledge Graphs (CKGs), methods like KICGPT [36] and FireRobBrain [37] have demonstrated significant improvements in knowledge completion tasks by leveraging structured knowledge prompts. However, these methods are susceptible to information loss and semantic bias during the conversion of triplets into textual prompts. In Temporal Knowledge Graphs (TKGs), approaches such as GenTKGQA [54] and zrLLM [55] have effectively handled temporal reasoning tasks through well-designed prompts and embeddings, though their performance heavily relies on the quality and clarity of the prompts. For Multi-modal Knowledge Graphs (MMKGs), MPIKGC [60] and MML [61] have shown promise in integrating multi-modal information into LLMs, but their effectiveness is limited by the LLMs’ ability to accurately interpret multi-modal features. In Inductive Knowledge Graphs (IKGs), methods like the fusion inductive framework [64] and LLAMA3 [65] have effectively guided LLMs in inductive reasoning tasks, but their success is contingent on high-quality, domain-specific training data. Open Knowledge Graphs (OKGs) have benefited from methods like UrbanKGent [70] and LLM-TIKG [76], which have enhanced reasoning abilities through fine-tuning, though the risk of generating hallucinations remains a concern. Lastly, Domain Knowledge Graphs (DKGs) have seen improvements in tasks such as medical question answering with KG-Rank [77] and biomarker discovery with KG-LLM [78], but their effectiveness is limited by the quality of domain-specific data and the risk of overfitting.

3.7.2. Scalability

The scalability of LLMs in different KG scenarios is influenced by several factors. In CKGs, fine-tuning methods like DIFT [42] and RecIn-Dial [43] are computationally intensive and require large amounts of high-quality annotated data, posing challenges for scalability. TKGs have seen more scalable approaches with prompt-based methods like GenTKGQA [54] and zrLLM [55], which do not require extensive fine-tuning but may struggle with complex temporal patterns. MMKGs face scalability challenges due to the complexity of multi-modal data and the need for high-quality, multi-modal training data. IKGs have benefited from prompt-based methods like the fusion inductive framework [64] and LLAMA3 [65], which are relatively scalable but require significant domain expertise for prompt design. OKGs have seen scalable solutions with fine-tuning methods like UrbanKGent [70] and LLM-TIKG [76], though the risk of overfitting remains a concern. DKGs have shown promise with fine-tuning methods like KG-Rank [77] and KG-LLM [78], but their scalability is limited by the need for domain-specific training data and the computational cost of fine-tuning large models.

3.7.3. Practical Impact

The practical impact of LLMs in KG tasks is evident across various domains. In CKGs, methods like KICGPT [36] and FireRobBrain [37] have enhanced knowledge completion and reasoning, though they may generate inaccurate results due to semantic bias. TKGs have benefited from methods like GenTKGQA [54] and zrLLM [55], which have improved temporal reasoning but may produce unreliable results if prompts are poorly designed. MMKGs have seen practical applications with MPIKGC [60] and MML [61], which have enhanced multi-modal knowledge representation but may struggle with certain modalities like images. IKGs have benefited from methods like the fusion inductive framework [64] and LLAMA3 [65], which have improved inductive reasoning but may generate hallucinations. OKGs have seen practical applications with UrbanKGent [70] and LLM-TIKG [76], which have enhanced reasoning abilities but may produce unreliable results due to data quality issues. DKGs have benefited from methods like KG-Rank [77] and KG-LLM [78], which have improved domain-specific knowledge management but may generate inaccurate results due to overfitting.

3.8. Analysis of LLMs in Tasks

3.8.1. Specific Improvements

The integration of LLMs into various types of Knowledge Graphs (KGs) has brought several specific improvements. In CKGs, LLMs have enhanced the accuracy of knowledge completion and reasoning tasks through their powerful text-generation capabilities. In TKGs, LLMs have improved temporal reasoning by capturing the semantic information in textual descriptions of temporal relations. MMKGs have benefited from LLMs’ ability to integrate multi-modal information, leading to more sophisticated knowledge representation. IKGs have seen improvements in inductive reasoning and knowledge transfer through LLMs’ ability to generalize from source to target graphs. OKGs have benefited from LLMs’ ability to handle open-domain knowledge and generate accurate and relevant information. DKGs have seen enhancements in domain-specific knowledge management through LLMs’ ability to generate human-like text and perform complex reasoning.

3.8.2. Defects

Despite these improvements, LLMs exhibit notable defects that pose challenges in different KG scenarios. A common issue is the “hallucination” phenomenon, where LLMs generate plausible but factually incorrect information. This is particularly problematic in critical applications such as medical knowledge graphs and cybersecurity, where accuracy is paramount. The black box nature of LLMs further complicates matters, as it makes their reasoning process difficult to interpret. This lack of transparency can undermine trust and accountability, especially in knowledge-intensive tasks that require high precision and reliability. Additionally, LLMs’ reliance on large amounts of high-quality training data and the risk of overfitting pose scalability and generalization challenges, especially in dynamic and evolving knowledge domains like OKGs and DKGs.

4. Prospective Directions

At present, in various KGE application scenarios, LLMs have already provided effective solutions to specific problems through different degrees of invocation. There are still many prospective research directions for the application of LLMs in various KGE scenarios in the future, which we will introduce in this section for researchers’ reference. We sincerely hope that the proposed potential directions can pique the research interests of beginners in this field. In the process of exploration, we have also found some examples in specific processes, which we will incorporate in the discussion of this section to facilitate comprehensibility for the readers.

4.1. Cross-Domain CKG Fusion

An investigation reveals that current applications of LLMs in the CKG domain primarily focus on single-domain KGs or generalized KGs, with limited exploration of cross-domain KG content fusion. For instance, studies have leveraged KGs and ranking techniques to enhance medical question answering in the medical domain [77] or empowered LLMs as political experts using KGs [38]. These works remain confined to single domains (e.g., medicine or politics) and overlook the synergistic potential with LLMs, leading to insufficient generalizability. Future research could utilize LLM comprehension and reasoning capabilities to align KG concepts across two domains by leveraging contextual and background knowledge—for example, identifying different nodes representing the same entity in two KGs.
Here we present a specific challenge for readers to contemplate: While LLMs are predominantly applied in single-domain KGs, cross-domain KG fusion confronts the challenge of concept alignment—such as entity ambiguity between medical and chemical domains. Cross-domain datasets like YAGO [31] and UMLS [90] can be leveraged to design a prompt-based bidirectional alignment mechanism. Specifically, LLMs can generate domain mapping rules (e.g., “Aspirin is an antipyretic drug in medicine and an organic compound in chemistry”), which are combined with the RotatE [91] model to optimize the cross-domain embedding space. The Hits@10 metric can be used to compare cross-domain link prediction accuracy, with alignment effects validated on the cross-domain subset of FB15k-237 [92].

4.2. Few-Shot Learning for TKG

Currently, the application aspects of LLMs for TKG are mainly focused on link prediction tasks or knowledge quiz tasks with sufficient samples, and very few articles apply LLMs to few-shot scenarios for TKG. Current research proposes [93] a novel few-shot solution using KG-driven incremental broad learning, and it leverages [76] ChatGPT’s few-shot learning capabilities to achieve data annotation and augmentation. However, the current work has not devoted effort to the processing of temporal information. The few existing methods attempt to augment the idea of generating rich relational descriptions by LLMs, but they neglect the use of temporal attributes for TKG. We suggest that in the future, LLMs can be fine-tuned on a small amount of TKG data to allow the model to capture temporal relations more accurately by adding explicit representations of temporal data. With prompt learning, LLMs can quickly adapt to few-shot scenarios and provide reasonable modeling of timing data.
Traditional TKG models suffer from insufficient generalization in scenarios with scarce samples (such as emerging infectious disease research), and LLMs struggle to capture the dynamic patterns of time series. A few-shot learning framework can be constructed based on the sparse time slices of the ICEWS 05-15 dataset [94], where LLMs generate natural language descriptions of temporal relations (e.g., “The efficacy improvement of Drug A in Q1 2020 is associated with viral mutations”), combined with meta-learning algorithms to fine-tune model parameters. It is recommended to measure the improvement of the MRR metric on the subset of the GDELT dataset [95].

4.3. Application of MMKG Tasks for More Specific Domains

The research of LLMs in domain-specific MMKG is still seriously insufficient, and this has only been attempted in a few specific domains. For example, ref. [76] leverages large language models and multimodal KGs to detect phishing web pages with or without logos. It lacks a targeted design for more domain-specific KGs involving multi-modal problems. LLMs can be useful for cross-modal information fusion to obtain information from multiple perspectives, providing richer and multidimensional semantic understanding. We propose that in future research, LLMs can be used to reason and fuse more domain-specific MMKG tasks based on the relations between text and other modalities (e.g., image, audio, etc.). For example, in the medical field, LLMs are able to combine the analysis results of CT scan images with the text of the patient’s medical history to provide more accurate disease diagnosis and personalized treatment plans.
In medical MMKGs, there exists a semantic gap in cross-modal alignment between text and medical images (e.g., mismatch between “pulmonary CT shadow” and medical terminology in text descriptions). It is proposed to adopt text-image pair data, use LLMs to generate semantic annotations for images (e.g., converting CT images into text descriptions like “ground-glass opacity in the right upper lobe of the lung”), and construct a cross-modal contrastive loss function in conjunction with the CLIP model to optimize the consistency of entity embeddings. The evaluation can be conducted on the MIMIC-CXR dataset by assessing the accuracy of MMKG completion with Precision@5.

4.4. Generic Framework Design for IKG

Since IKGs are relatively new concepts, the current applications of LLMs in IKG scenarios are still extremely underdeveloped, lacking a generic framework design for tasks such as link prediction in this direction. The survey revealed only a limited number of studies based on IKG, such as leveraging current-generation LLMs to achieve new entity link prediction in KGs through prompt generation, answer processing, entity candidate search, and final link prediction [65]. Both KG scenarios involve emergent new knowledge, and LLMs can play an important role in the process of updating new knowledge through their powerful language comprehension and reasoning capabilities. We propose using LLMs to integrate new knowledge from multiple sources, deal with redundant, conflicting, or missing information, and ensure the automatic correction and generation of new knowledge. For example, when confronted with inconsistent descriptions in different texts, the model can determine the most accurate knowledge based on context, timestamp, and other information.
Existing IKG approaches lack generalization ability for dynamically added new entities, and LLMs are prone to hallucinations when predicting unknown relations. We propose designing an incremental learning scenario based on the WN18RR [96] dataset, leveraging the in-context learning capability of LLMs to generate entity descriptions (e.g., “The relationship of the new entity ‘AI drug’ should be associated with ‘targeted therapy’”), and combining graph neural networks for inductive embedding updates. In the above scenario, the effectiveness can be validated by comparing the changes in AUC-ROC curves of link prediction under the condition of dynamically adding 20% new entities.

4.5. OKG Canonicalization

The redundancy and ambiguity characteristics of OKG can severely limit its related tasks, and the current LLMs are more often applied to the downstream tasks related to OKG, while the role of its expression canonicalization process is neglected. We observe that LLMs have been applied in OKG tasks such as electronic medical record [71], achieving data annotation and augmentation [76] and constructing KGs [72], but their application in the canonicalization tasks of OKG has not yet been identified. We have not found current research that leverages the rich pre-trained background of LLMs to assist in the canonicalization process. We propose obtaining richer knowledge background information based on LLMs and enhancing the soft clustering process for synonymous noun or verb phrase representations based on LLMs, which can help to fully utilize the sufficient pre-training advantages of LLMs.
In OKGs, there are a large number of redundant entities (e.g., “New York” and “Big Apple”), and LLMs have difficulty automatically identifying semantic equivalence. We propose using the COMBO dataset [97], generating semantic similarity scores for entity aliases through LLMs (such as calculating the semantic distance between “aspirin” and “acetylsalicylic acid”), and performing normalization in combination with K-means clustering. Subsequently, the entity alignment accuracy before and after normalization can be measured on the open subset of DBpedia [98].

4.6. Unified Representation Learning Framework for Multiple Types and Domains

After investigation, it is found that LLMs in different KG scenarios are often designed based on individual KG task scenarios or task scenarios in individual task domains, but there is a lack of a unified representation learning framework based on LLMs for multiple types and domains. It is recommended that the powerful language comprehension and highly scalable nature of LLMs be fully utilized in future research to establish a universal framework suitable for various task types, improve task generalization, and facilitate the development and integration of various types of KG downstream tasks.
The embedding spaces of KGs in different domains (such as biomedicine and social networks) are difficult to unify, and knowledge forgetting occurs in cross-domain migration of LLMs. We propose integrating the FB15k-237 [92] general dataset and the LMKG medical dataset [29] and designing an LLM fine-tuning framework based on adapters—setting up exclusive parameter layers for each domain and dynamically activating domain knowledge through prompt words (e.g., “Switch to biomedicine mode: explain drug-disease relationships”). In cross-domain link prediction tasks, the maximum mean difference (MMD) index of domain adapters is used to evaluate the consistency of embedding spaces.

4.7. Standard Framework and Criteria for Evaluating the Fusion Effect of LLM and KGE

The evaluation of the fusion effect between LLMs and KGE should be constructed based on the following three dimensions: knowledge representation ability, reasoning generalizability, and scenario adaptability. In terms of knowledge representation, the semantic consistency of entity embeddings (e.g., calculating the cosine similarity of cross-modal entity vectors) and the accuracy of relationship prediction (using traditional KGE metrics such as Hits@k and MRR) can be measured to verify whether the natural language prompts generated by LLMs effectively enhance the semantic expression capability of KGs. For reasoning generalizability, cross-domain migration tasks (such as migrating from general knowledge graphs to biomedical KGs) should be designed, and the AUC-ROC curve of link prediction before and after migration should be compared to evaluate the model’s knowledge generalization ability in different scenarios.
For temporal knowledge graphs, the focus should be on examining the model’s ability to capture dynamic temporal features, which can be verified by measuring the change in prediction error after dynamically adding timestamps on temporal datasets such as ICEWS. For multimodal KGs, cross-modal alignment metrics (such as the text-image matching score of the CLIP model) need to be introduced to evaluate whether LLMs effectively bridge the semantic gap between modalities such as text and images. In addition, a robustness evaluation mechanism should be established to test the model’s anti-interference ability in knowledge pollution scenarios by injecting adversarial perturbations, ensuring the practical application reliability of the fusion method.

4.8. Summary of the Advantages and Disadvantages of Existing Research Work

In recent years, significant progress has been made in the integration of LLM for KGE, demonstrating multiple advantages. Firstly, LLM significantly improves the performance of KGE tasks through its powerful semantic understanding and generation capabilities. For example, LLM can transform structured knowledge into natural language descriptions through prompt engineering in CKG, thereby enhancing the accuracy of knowledge reasoning. In TKG, LLM is capable of processing dynamic temporal information, supporting event prediction and drug efficacy analysis. In addition, MMKG achieves the fusion of text and image modalities through LLM, further enriching knowledge representation. The introduction of LLM also addresses the limitations of traditional KGE methods in handling long tail entities, dynamic knowledge updates, and cross domain inference, improving the model’s generalization ability.
However, there are still some shortcomings in existing research. On the one hand, the “illusion” problem of LLM (i.e., generating seemingly reasonable but actually erroneous knowledge) may have serious consequences in critical areas such as healthcare and cybersecurity. On the other hand, the black box nature of LLM makes its inference process difficult to explain, which affects the credibility of the model. In addition, existing methods are often designed for a single task or domain, lacking a unified framework to support KGE tasks of multiple types and domains. For example, the canonicalization of cross domain knowledge graph fusion and OKG is still in the exploratory stage. Data quality and annotation cost are also key factors that constrain the deep integration of LLM and KGE, especially in scenarios with few samples and low resources. Future research needs to further optimize the robustness and interpretability of the model and explore more efficient cross domain and dynamic knowledge integration methods.

5. Datasets and Code Resources

In this section, we meticulously curate a collection of classic datasets applicable to various KGE scenarios, aiming to provide valuable resources for readers. Additionally, we systematically organize the code resources derived from some of the articles cited in this paper, facilitating a more comprehensive understanding and practical application for the readers.
It should be noted that the download links for the following datasets are all presented in Table 4.

5.1. Dataset

In this section, we introduce the various related datasets and have listed the relevant metrics for the comparison of the different datasets in Table 5.

5.1.1. ICEWS 05-15

ICEWS is an integrated conflict early warning system, with data sourced from global news media, government reports, and sociological research. It contains over 5 millions social and political events from 2005 to 2015. ICEWS 05-15 includes about 50 k entities (such as individuals, organizations, countries, etc.) and approximately 4 millions event triplets. It records political, social, and economic events on a global scale, suitable for time-series KGs and event prediction tasks.

5.1.2. GDELT

GDELT is a global event database developed by Google, with data sourced from global news media reports. It contains over 3 billion event records. GDELT includes about 100 k entities (such as countries, cities, organizations, etc.) and approximately 3 k millions event triplets. It records events, language, and emotional information on a global scale, suitable for social science research such as conflict analysis and political stability research.

5.1.3. YAGO

YAGO is a KG extracted from resources such as Wikipedia, WordNet, and GeoNames. It contains over 1 million entities and over 350 k facts. YAGO includes about 1.2 millions entities (such as names, locations, events, etc.) and approximately 350 k triplets. It provides rich semantic information, suitable for tasks such as entity linking, relation extraction, and KG construction.

5.1.4. UMLS

UMLS is a comprehensive medical terminology system developed by the National Library of Medicine in the United States. It contains over 300 k medical terms and concepts. UMLS includes about 300 k medical terms (such as diseases, drugs, treatment methods, etc.) and approximately 1 million triplets. It is used for biomedical text processing, information retrieval, and KG construction.

5.1.5. MetaQA

MetaQA is a multi hop question answering dataset for question answering systems, sourced from a movie database. It includes 1-hop, 2-hop, and 3-hop question answering pairs. It contains approximately 4.3k entities (such as movie characters, movie titles, etc.) and about 300 k triplets. MetaQA is suitable for machine reading comprehension, question answering system development, and multi-hop reasoning tasks.

5.1.6. ReVerb45K

ReVerb45K is a set of relation facts automatically extracted from web page text. It contains 4.5 k relation instances, about 10 k entities, and about 45 k triplets. ReVerb45K is used for relation extraction and KG construction tasks.

5.1.7. FB15k-237

FB15k-237 is a subset of the Freebase KG that has been processed to reduce overlap between the training and testing sets. It contains 15028 entities, 237 relations, and approximately 310 k triplets. FB15k-237 is widely used for KG completion and link prediction tasks.

5.1.8. WN18RR

WN18RR is a subset of the WordNet KG that has been revised to increase task difficulty. It contains 40,943 entities, 11 types of relations, 40,943 entities, and approximately 93 k triplets. WN18RR is mainly used for research on KG completion and link prediction.

5.2. Code Resources

Table 6 lists some important code resources and open source projects related to the research work discussed in this article. These resources cover multiple cutting-edge research directions combining large-scale pre-trained language models with KG. Through these open-source codes, readers can gain a deeper understanding of the implementation details of related research, reproduce experimental results, or make further extensions and innovations.

6. Conclusions

LLMs have been widely used in tasks in multiple scenarios with KGEs [93,99,100,101,102]. Currently, this promising field is attracting increasing attention from researchers. A large number of advanced methods has been proposed. In this paper, we propose a new taxonomy for the investigation of the current applications of LLMs in the field of KGE. Specifically, we categorize them according to the KG scenarios in which they are applied and innovate them according to the different degrees of use for LLMs in different scenarios. We not only discuss the use of LLMs in various KGE tasks, but we also explore the relevant applications and innovation directions that may be developed in the future. Through this survey, we offer a comprehensive and systematic overview of the application of LLMs within the KGE domain. We sincerely hope that this overview can serve as a valuable guide for future research endeavors.

Funding

This research is sponsored by the Key R&D Program of Shandong Province, China, (No. 2023CXGC010801) and the National Key R&D Program of China (2019YFA0709401).

Data Availability Statement

No new data were created or analyzed in this study.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Vidal, M.; Chudasama, Y.; Huang, H.; Purohit, D.; Torrente, M. Poisoning scientific knowledge using large language models. J. Web Semant. 2025, 84. [Google Scholar]
  2. Gao, F.; He, Q.; Wang, X.; Qiu, L.; Huang, M. An Efficient Rumor Suppression Approach with Knowledge Graph Convolutional Network in Social Network. IEEE Trans. Comput. Soc. Syst. 2024, 11, 6254–6267. [Google Scholar] [CrossRef]
  3. Yang, Y.; Peng, C.; Cao, E.-Z.; Zou, W. Building Resilience in Supply Chains: A Knowledge Graph-Based Risk Management Framework. IEEE Trans. Comput. Soc. Syst. 2024, 11, 3873–3881. [Google Scholar] [CrossRef]
  4. Jiang, C.; Tang, M.; Jin, S.; Huang, W.; Liu, X. KGNMDA: A Knowledge Graph Neural Network Method for Predicting Microbe-Disease Associations. IEEE/ACM Trans. Comput. Biol. Bioinform. 2023, 20, 1147–1155. [Google Scholar] [CrossRef] [PubMed]
  5. Zhao, L.; Qi, X.; Chen, Y.; Qiao, Y.; Bu, D.; Wu, Y.; Luo, Y.; Wang, S.; Zhang, R.; Zhao, Y. Biological Knowledge Graph-Guided Investigation of Immune Therapy Response in Cancer with Graph Neural Network. Brief. Bioinform. 2023, 24, bbad023. [Google Scholar] [CrossRef]
  6. Huang, C.; Chen, D.; Fan, T.; Wu, B.; Yan, X. Incorporating Environmental Knowledge Embedding and Spatial-Temporal Graph Attention Networks for Inland Vessel Traffic Flow Prediction. Eng. Appl. Artif. Intell. 2024, 133, 108301. [Google Scholar] [CrossRef]
  7. Mlodzian, L.; Sun, Z.; Berkemeyer, H.; Monka, S.; Wang, Z.; Dietze, S.; Halilaj, L.; Luettin, J. nuScenes Knowledge Graph—A Comprehensive Semantic Representation of Traffic Scenes for Trajectory Prediction. In Proceedings of the 2023 IEEE/CVF International Conference on Computer Vision Workshops (ICCVW), Paris, France, 2–6 October 2023; pp. 42–52. [Google Scholar] [CrossRef]
  8. Agrawal, G.; Pal, K.; Deng, Y.; Liu, H.; Chen, Y.-C. CyberQ: Generating Questions and Answers for Cybersecurity Education Using Knowledge Graph-Augmented LLMs. In Proceedings of the AAAI Conference on Artificial Intelligence 2024, Vancouver, BC, Canada, 20–27 February 2024; Volume 38, pp. 23164–23172. [Google Scholar] [CrossRef]
  9. Panda, P.; Agarwal, A.; Devaguptapu, C.; Kaul, M.; Ap, P. HOLMES: Hyper-Relational Knowledge Graphs for Multi-hop Question Answering Using LLMs. In Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Bangkok, Thailand, 11–16 August 2024; pp. 13263–13282. [Google Scholar]
  10. Sun, L.; Tao, Z.; Li, Y.; Arakawa, H. ODA: Observation-Driven Agent for Integrating LLMs and Knowledge Graphs. In Proceedings of the Findings of the Association for Computational Linguistics: ACL 2024, Bangkok, Thailand, 11–16 August 2024; Association for Computational Linguistics: Kerrville, TX, USA, 2024; pp. 7417–7431. [Google Scholar]
  11. Venkatakrishnan, R.; Tanyildizi, E.; Canbaz, M.A. Semantic interlinking of Immigration Data using LLMs for Knowledge Graph Construction. In Proceedings of the ACM Web Conference 2024 (WWW ’24), Singapore, 13–17 May 2024; Association for Computing Machinery: New York, NY, USA, 2024; pp. 605–608. [Google Scholar] [CrossRef]
  12. Sun, L.; Zhang, P.; Gao, F.; An, Y.; Li, Z.; Zhao, Y. SF-GPT: A training-free method to enhance capabilities for knowledge graph construction in LLMs. Neurocomputing 2025, 613, 128726. [Google Scholar] [CrossRef]
  13. Pan, S.; Luo, L.; Wang, Y.; Chen, C.; Wang, J.; Wu, X. Unifying Large Language Models and Knowledge Graphs: A Roadmap. IEEE Trans. Knowl. Data Eng. 2024, 36, 3580–3599. [Google Scholar] [CrossRef]
  14. Yang, L.; Chen, H.; Li, Z.; Ding, X.; Wu, X. Give us the Facts: Enhancing Large Language Models with Knowledge Graphs for Fact-Aware Language Modeling. IEEE Trans. Knowl. Data Eng. 2024, 36, 3091–3110. [Google Scholar] [CrossRef]
  15. Liu, L.; Xu, P.; Fan, K.; Wang, M. Research on application of knowledge graph in industrial control system security situation awareness and decision-making: A survey. Neurocomputing 2025, 613, 128721. [Google Scholar] [CrossRef]
  16. Zhu, B.; Wang, R.; Wang, J.; Zhao, F.; Wang, K. A Survey: Knowledge Graph Entity Alignment Research Based on Graph Embedding. Artif. Intell. Rev. 2024, 57, 229. [Google Scholar] [CrossRef]
  17. Huang, X.; Ruan, W.; Huang, W.; Jin, G.; Dong, Y.; Wu, C.; Bensalem, S.; Mu, R.; Qi, Y.; Zhao, X.; et al. A survey of safety and trustworthiness of large language models through the lens of verification and validation. Artif. Intell. Rev. 2024, 57, 175. [Google Scholar] [CrossRef]
  18. Kumar, P. Large language models (LLMs): Survey, technical frameworks, and future challenges. Artif. Intell. Rev. 2024, 57, 260. [Google Scholar] [CrossRef]
  19. Zhao, S.; Sun, X. Enabling controllable table-to-text generation via prompting large language models with guided planning. Knowl.-Based Syst. 2024, 304, 112571. [Google Scholar] [CrossRef]
  20. Cui, M.; Du, J.; Zhu, S.; Xiong, D. Efficiently Exploring Large Language Models for Document-Level Machine Translation with In-context Learning. In Proceedings of the Findings of the Association for Computational Linguistics: ACL 2024, Bangkok, Thailand, 11–16 August 2024; Association for Computational Linguistics: Kerrville, TX, USA, 2024; pp. 10885–10897. [Google Scholar]
  21. Liu, C.; Wang, C.; Peng, Y.; Li, Z. ZVQAF: Zero-shot visual question answering with feedback from large language models. Neurocomputing 2024, 580, 127505. [Google Scholar] [CrossRef]
  22. Hellwig, N.C.; Fehle, J.; Wolff, C. Exploring large language models for the generation of synthetic training samples for aspect-based sentiment analysis in low resource settings. Expert Syst. Appl. 2025, 261, 125514. [Google Scholar] [CrossRef]
  23. Song, H.; Su, H.; Shalyminov, I.; Cai, J.; Mansour, S. FineSurE: Fine-grained Summarization Evaluation using LLMs. In Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Bangkok, Thailand, 11–16 August 2024; Association for Computational Linguistics: Kerrville, TX, USA, 2024; pp. 906–922. [Google Scholar]
  24. Zhou, Z.; Elejalde, E. Unveiling the silent majority: Stance detection and characterization of passive users on social media using collaborative filtering and graph convolutional networks. EPJ Data Sci. 2024, 13, 28. [Google Scholar] [CrossRef]
  25. Cui, W.; Shang, M. KAGN: Knowledge-powered attention and graph convolutional networks for social media rumor detection. J. Big Data 2023, 10, 45. [Google Scholar] [CrossRef]
  26. Huang, Q.; Chen, J. Enhancing academic performance prediction with temporal graph networks for massive open online courses. J. Big Data 2024, 11, 52. [Google Scholar] [CrossRef]
  27. Teji, B.; Roy, S.; Dhami, D.S.; Bhandari, D.; Guzzi, P.H. Graph Embedding Techniques for Predicting Missing Links in Biological Networks: An Empirical Evaluation. IEEE Trans. Emerg. Top. Comput. 2024, 12, 190–201. [Google Scholar] [CrossRef]
  28. Ying, H.; Zhao, Z.; Zhao, Y.; Zeng, S.; Yu, S. CoRTEx: Contrastive learning for representing terms via explanations with applications on constructing biomedical knowledge graphs. J. Am. Med. Inform. Assoc. 2024, 31, 1912–1920. [Google Scholar] [CrossRef] [PubMed]
  29. Yang, P.; Wang, H.; Huang, Y.; Yang, S.; Zhang, Y.; Huang, L.; Zhang, Y.; Wang, G.; Yang, S.; He, L.; et al. LMKG: A large-scale and multi-source medical knowledge graph for intelligent medicine applications. Knowl.-Based Syst. 2024, 284, 111323. [Google Scholar] [CrossRef]
  30. Liu, B.; Hou, S.; Zhong, W.; Zhao, X.; Liu, Y.; Yang, Y. Enhancing Temporal Knowledge Graph Alignment in News Domain with Box Embedding. IEEE Trans. Comput. Soc. Syst. 2024, 11, 4909–4919. [Google Scholar] [CrossRef]
  31. Suchanek, F.M.; Kasneci, G.; Weikum, G. Yago: A core of semantic knowledge. In Proceedings of the 16th International Conference on World Wide Web, Banff, AB, Canada, 8–12 May 2007; pp. 697–706. [Google Scholar]
  32. Singhal, A. Introducing the Knowledge Graph: Things, Not Strings. Official Google Blog. 16 May 2012. Available online: https://blog.google/products/search/introducing-knowledge-graph-things-not/ (accessed on 2 January 2015).
  33. Young, M. The Technical Writer’s Handbook; University Science: Mill Valley, CA, USA, 1989. [Google Scholar]
  34. Jiang, C.; Jiang, Y.; Wu, W.; Zheng, Y.; Xie, P.; Tu, K. COMBO: A Complete Benchmark for Open KG Canonicalization. In Proceedings of the 17th Conference of the European Chapter of the Association for Computational Linguistics, Dubrovnik, Croatia, 2–6 May 2023; Association for Computational Linguistics: Kerrville, TX, USA, 2023; pp. 340–357. [Google Scholar]
  35. Cao, Z.; Luo, C. Link prediction for knowledge graphs based on extended relational graph attention networks. Expert Syst. Appl. 2025, 259, 125260. [Google Scholar] [CrossRef]
  36. Wei, Y.; Huang, Q.; Zhang, Y.; Kwok, J. KICGPT: Large Language Model with Knowledge in Context for Knowledge Graph Completion. In Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2023, Singapore, 6–10 December 2023; Association for Computational Linguistics: Kerrville, TX, USA, 2023; pp. 8667–8683. [Google Scholar]
  37. Zhang, J.; Cai, S.; Jiang, Z.; Xiao, J.; Ming, Z. FireRobBrain: Planning for a Firefighting Robot Using Knowledge Graph and Large Language Model. In Proceedings of the 2024 10th IEEE International Conference on Intelligent Data and Security (IDS), New York, NY, USA, 6–9 May 2024; pp. 37–41. [Google Scholar] [CrossRef]
  38. Mou, X.; Li, Z.; Lyu, H.; Luo, J.; Wei, Z. Unifying Local and Global Knowledge: Empowering Large Language Models as Political Experts with Knowledge Graphs. In Proceedings of the ACM Web Conference 2024 (WWW ’24), Singapore, 13–17 May 2024; Association for Computing Machinery: New York, NY, USA, 2024; pp. 2603–2614. [Google Scholar] [CrossRef]
  39. Wang, Y.; Hu, M.; Huang, Z.; Li, D.; Yang, D.; Lu, X. KC-GenRe: A Knowledge-constrained Generative Re-ranking Method Based on Large Language Models for Knowledge Graph Completion. In Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024), Torin, Italy, 20–25 May 2024; pp. 9668–9680. [Google Scholar]
  40. Yang, R.; Zhu, J.; Man, J.; Fang, L.; Zhou, Y. Enhancing text-based knowledge graph completion with zero-shot large language models: A focus on semantic enhancement. Knowl.-Based Syst. 2024, 300, 112155. [Google Scholar] [CrossRef]
  41. Hou, W.; Zhao, W.; Jia, N.; Liu, X. Low-resource knowledge graph completion based on knowledge distillation driven by large language models. Appl. Soft Comput. 2025, 169, 112622. [Google Scholar] [CrossRef]
  42. Liu, Y.; Tian, X.; Sun, Z.; Hu, W. Finetuning Generative Large Language Models with Discrimination Instructions for Knowledge Graph Completion. In Proceedings of the Semantic Web–ISWC 2024, Hanover, MD, USA, 11 November 2024; Demartini, G., Hose, K., Acosta, M., Palmonari, M., Cheng, G., Skaf-Molli, H., Ferranti, N., Hernández, D., Hogan, A., Eds.; Lecture Notes in Computer Science. Springer: Cham, Switzerland, 2025; Volume 15231. [Google Scholar]
  43. Wang, L.; Hu, H.; Sha, L.; Xu, C.; Jiang, D.; Wong, K.-F. RecInDial: A Unified Framework for Conversational Recommendation with Pretrained Language Models. In Proceedings of the 2nd Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics and the 12th International Joint Conference on Natural Language Processing, Online, 21–23 November 2022; Association for Computational Linguistics: Kerrville, TX, USA, 2022; Volume 1, pp. 489–500. [Google Scholar]
  44. Shu, D.; Chen, T.; Jin, M.; Zhang, C.; Du, M.; Zhang, Y. Knowledge Graph Large Language Model (KG-LLM) for Link Prediction. arXiv 2024, arXiv:2403.07311. [Google Scholar]
  45. Alqaaidi, S.K.; Kochut, K. Relations Prediction for Knowledge Graph Completion Using Large Language Models. arXiv 2024, arXiv:2405.02738. [Google Scholar]
  46. Tian, S.; Luo, Y.; Xu, T.; Yuan, C.; Jiang, H.; Wei, C.; Wang, X. KG-Adapter: Enabling Knowledge Graph Integration in Large Language Models through Parameter-Efficient Fine-Tuning. In Proceedings of the Findings of the Association for Computational Linguistics: ACL 2024, Bangkok, Thailand, 11–16 August 2024; Association for Computational Linguistics: Kerrville, TX, USA, 2024; pp. 3813–3828. [Google Scholar]
  47. Zhang, Z.; Gu, J.; Zhou, J.; Pan, J.Z.; Zhang, W.; Chen, H. Learning to Plan for Retrieval-Augmented Large Language Models from Knowledge Graphs. In Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2024, Miami, FL, USA, 7–11 December 2024; Association for Computational Linguistics: Kerrville, TX, USA, 2024; pp. 7813–7835. [Google Scholar]
  48. Bianchini, F.; Calamo, M.; De Luzi, F.; Macrì, M.; Mecella, M. Enhancing Complex Linguistic Tasks Resolution Through Fine-Tuning LLMs, RAG and Knowledge Graphs (Short Paper). In Proceedings of the Advanced Information Systems Engineering Workshops, CAiSE 2024, Limassol, Cyprus, 3–4 June 2024; Almeida, J.P.A., Di Ciccio, C., Kalloniatis, C., Eds.; Lecture Notes in Business Information Processing. Springer: Cham, Switzerland, 2024; Volume 521, pp. 176–181. [Google Scholar] [CrossRef]
  49. Hu, S.; Zou, G.; Yang, S.; Gan, Y.; Zhang, B.; Chen, Y. Large Language Model Meets Graph Neural Network in Knowledge Distillation. arXiv 2024, arXiv:2402.05894. [Google Scholar] [CrossRef]
  50. Zhang, M.; Ye, X.; Liu, Q.; Ren, P.; Wu, S.; Chen, Z. Knowledge Graph Enhanced Large Language Model Editing. In Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, Miami, Florida, USA, 7–11 December 2024; Association for Computational Linguistics: Kerrville, TX, USA, 2024; pp. 22647–22662. [Google Scholar]
  51. Salnikov, M.; Le, H.; Rajput, P.; Nikishina, I.; Braslavski, P.; Malykh, V.; Panchenko, A. Large Language Models Meet Knowledge Graphs to Answer Factoid Questions. In Proceedings of the 37th Pacific Asia Conference on Language, Information and Computation, Hong Kong, China, 12–14 December 2023; Association for Computational Linguistics: Kerrville, TX, USA, 2023; pp. 635–644. [Google Scholar]
  52. Geng, Y.; Chen, J.; Zeng, Y.; Chen, Z.; Zhang, W.; Pan, J.Z.; Wang, Y.; Xu, X. Prompting disentangled embeddings for knowledge graph completion with pre-trained language model. Expert Syst. Appl. 2025, 268, 126175. [Google Scholar] [CrossRef]
  53. Sun, J.; Xu, C.; Tang, L.; Wang, S.; Lin, C.; Gong, Y.; Ni, L.; Shum, H.-Y.; Guo, J. Think-on-Graph: Deep and Responsible Reasoning of Large Language Model on Knowledge Graph. In Proceedings of the Twelfth International Conference on Learning Representations, Vienna, Austria, 7–11 May 2024. [Google Scholar]
  54. Gao, Y.; Qiao, L.; Kan, Z.; Wen, Z.; He, Y.; Li, D. Two-stage Generative Question Answering on Temporal Knowledge Graph Using Large Language Models. In Proceedings of the Findings of the Association for Computational Linguistics: ACL 2024, Bangkok, Thailand, 11–16 August 2024; Association for Computational Linguistics: Kerrville, TX, USA, 2024; pp. 6719–6734. [Google Scholar]
  55. Ding, Z.; Cai, H.; Wu, J.; Ma, Y.; Liao, R.; Xiong, B.; Tresp, V. zrLLM: Zero-Shot Relational Learning on Temporal Knowledge Graphs with Large Language Models. In Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers), Mexico City, Mexico, 16–21 June 2024; Association for Computational Linguistics: Kerrville, TX, USA, 2024; pp. 1877–1895. [Google Scholar]
  56. Xia, Y.; Wang, D.; Liu, Q.; Wang, L.; Wu, S.; Zhang, X. Chain-of-History Reasoning for Temporal Knowledge Graph Forecasting. arXiv 2024, arXiv:2402.14382. [Google Scholar]
  57. Liao, R.; Jia, X.; Li, Y.; Ma, Y.; Tresp, V. GenTKG: Generative Forecasting on Temporal Knowledge Graph with Large Language Models. In Proceedings of the Findings of the Association for Computational Linguistics: NAACL 2024, Mexico City, Mexico, 16–21 June 2024; Association for Computational Linguistics: Kerrville, TX, USA, 2024; pp. 4303–4317. [Google Scholar]
  58. Li, Q.; Wu, G. Explainable Reasoning over Temporal Knowledge Graphs by Pre-trained Language Model. Inf. Process. Manag. 2025, 62, 103903. [Google Scholar] [CrossRef]
  59. Deng, Y.; Song, J.; Yang, Z.; Long, Y.; Zeng, L.; Zhou, L. Diachronic Semantic Encoding Based on Pre-trained Language Model for Temporal Knowledge Graph Reasoning. Knowl.-Based Syst. 2025, 318, 113479. [Google Scholar] [CrossRef]
  60. Xu, D.; Zhang, Z.; Lin, Z.; Wu, X.; Zhu, Z.; Xu, T.; Zhao, X.; Zheng, Y.; Chen, E. Multi-perspective Improvement of Knowledge Graph Completion with Large Language Models. In Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024), Torino, Italy, 22–24 May 2024; pp. 11956–11968. [Google Scholar]
  61. Yu, Z.; Wang, S.; Zhu, Y.; Zou, Z. A Multimodal Ecological Civilization Pattern Recommendation Method Based on Large Language Models and Knowledge Graph. arXiv 2023, arXiv:2310.18951. [Google Scholar]
  62. Li, Y.; Huang, C.; Deng, S.; Lock, M.L.; Cao, T.; Oo, N.; Lim, H.W.; Hooi, B. KnowPhish: Large Language Models Meet Multimodal Knowledge Graphs for Enhancing Reference-Based Phishing Detection. In Proceedings of the 33rd USENIX Security Symposium (USENIX Security 24), Philadelphia, PA, USA, 14–16 August 2024; USENIX Association: Berkeley, CA, USA, 2024; pp. 793–810. [Google Scholar]
  63. Sun, Q.; Luo, Y.; Zhang, W.; Li, S.; Li, J.; Niu, K.; Kong, X.; Liu, W. Docs2KG: Unified Knowledge Graph Construction from Heterogeneous Documents Assisted by Large Language Models. arXiv 2024, arXiv:2406.02962. [Google Scholar]
  64. Xing, X.; Jia, B.; Huang, Z.; Chen, Y.; Wang, J.; Fan, A.; Chen, X.; Cao, L. A Fusion Inference Method for Large Language Models and Knowledge Graphs Based on Structured Injection and Causal Inference. In Proceedings of the 2023 5th International Conference on Internet of Things, Automation and Artificial Intelligence (IoTAAI 2023), Nanchang, China, 24–26 November 2023; ACM: New York, NY, USA, 2023; pp. 208–213. [Google Scholar] [CrossRef]
  65. Hamann, F.; Falk, M.; Walker, L. Expanding Knowledge Graphs through Text: Leveraging Large Language Models for Inductive Link Prediction. In Proceedings of the INFORMATIK 2024, Wiesbaden, Germany, 24–26 September 2024; Gesellschaft für Informatik e.V.: Bonn, Germany, 2024; pp. 1407–1417. [Google Scholar]
  66. Zhao, Q.; Qian, H.; Liu, Z.; Zhang, G.-D.; Gu, L. Breaking the Barrier: Utilizing Large Language Models for Industrial Recommendation Systems through an Inferential Knowledge Graph. In Proceedings of the 33rd ACM International Conference on Information and Knowledge Management (CIKM 2024), Boise, ID, USA, 21–25 October 2024; ACM: New York, NY, USA, 2024; pp. 5086–5093. [Google Scholar] [CrossRef]
  67. Wang, K.; Xu, Y.; Wu, Z.; Luo, S. LLM as Prompter: Low-resource Inductive Reasoning on Arbitrary Knowledge Graphs. In Proceedings of the Findings of the Association for Computational Linguistics: ACL 2024, Bangkok, Thailand, 11–16 August 2024; ACL: Kerrville, TX, USA, 2024; pp. 3742–3759. [Google Scholar]
  68. Liu, J.; Zhan, J. Constructing Knowledge Graph from Cyber Threat Intelligence Using Large Language Model. In Proceedings of the 2023 IEEE International Conference on Big Data (BigData 2023), Sorrento, Italy, 15–18 December 2023; pp. 516–521. [Google Scholar] [CrossRef]
  69. Park, C.; Lee, H.; Jeong, O.-R. Leveraging Medical Knowledge Graphs and Large Language Models for Enhanced Mental Disorder Information Extraction. Future Internet 2024, 16, 260. [Google Scholar] [CrossRef]
  70. Ning, Y.; Liu, H. UrbanKGent: A Unified Large Language Model Agent Framework for Urban Knowledge Graph Construction. arXiv 2024, arXiv:2402.06861. [Google Scholar]
  71. Arsenyan, V.; Bughdaryan, S.; Shaya, F.; Small, K.W.; Shahnazaryan, D. Large Language Models for Biomedical Knowledge Graph Construction: Information Extraction from EMR Notes. In Proceedings of the 23rd Workshop on Biomedical Natural Language Processing, Bangkok, Thailand, 16 August 2024; ACL: Kerrville, TX, USA, 2024; pp. 295–317. [Google Scholar]
  72. Xu, T.; Gu, Y.; Xue, M.; Gu, R.; Li, B.; Gu, X. Knowledge Graph Construction for Heart Failure Using Large Language Models with Prompt Engineering. Front. Comput. Neurosci. 2024, 18, 1389475. [Google Scholar] [CrossRef]
  73. AlMahri, S.; Xu, L.; Brintrup, A. Enhancing Supply Chain Visibility with Knowledge Graphs and Large Language Models. arXiv 2024, arXiv:2408.07705. [Google Scholar]
  74. Datta, P.; Vitiugin, F.; Chizhikova, A.; Sawhney, N. Construction of Hyper-Relational Knowledge Graphs Using Pre-Trained Large Language Models. arXiv 2024, arXiv:2403.11786. [Google Scholar]
  75. Chen, H.; Méndez, S.J.R.; Omran, P.G. Open Local Knowledge Graph Construction from Academic Papers Using Generative Large Language Models. In Proceedings of the Companion Proceedings of the ACM on Web Conference 2025 (WWW ’25), Sydney, Australia, 28 April–2 May 2025; ACM: New York, NY, USA, 2025; pp. 2551–2559. [Google Scholar] [CrossRef]
  76. Hu, Y.; Zou, F.; Han, J.; Sun, X.; Wang, Y. LLM-TIKG: Threat Intelligence Knowledge Graph Construction Utilizing Large Language Model. Comput. Secur. 2024, 145, 103999. [Google Scholar] [CrossRef]
  77. Yang, R.; Liu, H.; Marrese-Taylor, E.; Zeng, Q.; Ke, Y.; Li, W.; Cheng, L.; Chen, Q.; Caverlee, J.; Matsuo, Y.; et al. KG-Rank: Enhancing Large Language Models for Medical QA with Knowledge Graphs and Ranking Techniques. In Proceedings of the 23rd Workshop on Biomedical Natural Language Processing, Bangkok, Thailand, 16 August 2024; Association for Computational Linguistics: Kerrville, TX, USA, 2024; pp. 155–166. [Google Scholar]
  78. Karim, R.; Molinas Comet, L.; Shajalal, M.; De Perthuis, P.; Rebholz-Schuhmann, D.; Decker, S. From Large Language Models to Knowledge Graphs for Biomarker Discovery in Cancer. In Proceedings of the 57th Hawaii International Conference on System Sciences, Honolulu, HI, USA, 3–6 January 2024; HICSS: Waikoloa, HI, USA, 2024; pp. 5576–5586. [Google Scholar]
  79. Du, J.; Yin, H. KLDP: A data profiling technique based on knowledge graph and large language modeling. In Proceedings of the 2023 IEEE 22nd International Conference on Trust, Security and Privacy in Computing and Communications (TrustCom), Exeter, UK, 10–12 July 2023; pp. 2333–2340. [Google Scholar] [CrossRef]
  80. Wang, C.; Hua, M.; Song, J.; Tang, X.-S. Knowledge graphs enhanced large language model prompt for electric power question answering. In Proceedings of the 2023 7th International Conference on Electronic Information Technology and Computer Engineering (EITCE ’23), Zhuhai, China, 27–29 October 2023; Association for Computing Machinery: New York, NY, USA, 2023; pp. 24–29. [Google Scholar] [CrossRef]
  81. Yang, J.; Xu, H.; Mirzoyan, S.; Chen, T.; Liu, Z.; Ju, W.; Liu, L.; Zhang, M.; Wang, S. Integrating knowledge graphs with symbolic AI: The path to interpretable hybrid AI systems in medicine. Web Semant. 2025, 84, 100856. [Google Scholar] [CrossRef]
  82. Pei, A.; Yang, Z.; Zhu, S.; Cheng, R.; Jia, J.; Wang, L. KGPA: Robustness evaluation for large language models via cross-domain knowledge graphs. arXiv 2024, arXiv:2406.10802. [Google Scholar]
  83. Kumar, N.M.; Lisa, F.T.; Islam, S.R. Prompt chaining-assisted malware detection: A hybrid approach utilizing fine-tuned LLMs and domain knowledge-enriched cybersecurity knowledge graphs. In Proceedings of the 2024 IEEE International Conference on Big Data (BigData), Washington, DC, USA, 4–7 December 2024; pp. 1672–1677. [Google Scholar] [CrossRef]
  84. Dernbach, S.; Agarwal, K.; Zuniga, A.; Henry, M.; Choudhury, S. GLaM: Fine-tuning large language models for domain knowledge graph alignment via neighborhood partitioning and generative subgraph encoding. In Proceedings of the AAAI Symposium Series 2024, Stanford, CA, USA, 25–27 March 2024; Volume 3, pp. 82–89. [Google Scholar] [CrossRef]
  85. Li, Y.; He, Y.; Lian, R.; Guo, Q. Fault diagnosis and system maintenance based on large language models and knowledge graphs. In Proceedings of the 2023 5th International Conference on Robotics, Intelligent Control and Artificial Intelligence (RICAI), Hangzhou, China, 20–23 November 2023; pp. 589–592. [Google Scholar] [CrossRef]
  86. Liu, P.; Qian, L.; Zhao, X.; Tao, B. Joint knowledge graph and large language model for fault diagnosis and its application in aviation assembly. IEEE Trans. Ind. Informat. 2024, 20, 8160–8169. [Google Scholar] [CrossRef]
  87. Wawrzik, F.; Plaue, M.; Vekariya, S.; Grimm, C. Customized Information and Domain-centric Knowledge Graph Construction with Large Language Models. arXiv 2024, arXiv:2409.20010. [Google Scholar]
  88. Qi, Z.; Yu, Y.; Tu, M.; Tan, J.; Huang, Y. FoodGPT: A Large Language Model in the Food Testing Domain with Incremental Pre-training and Knowledge Graph Prompt. arXiv 2023, arXiv:2308.10173. [Google Scholar]
  89. Xiao, Y.; Zhang, S.; Zhou, H.; Li, M.; Yang, H.; Zhang, R. FuseLinker: Leveraging LLM’s pre-trained text embeddings and domain knowledge to enhance GNN-based link prediction on biomedical knowledge graphs. J. Biomed. Inform. 2024, 158, 104730. [Google Scholar] [CrossRef] [PubMed]
  90. Bodenreider, O.; Burgun, A. Aligning Knowledge Sources in the UMLS: Methods, Quantitative Results, and Applications. In MEDINFO 2004, Studies in Health Technology and Informatics; IOS Press: Amsterdam, The Netherlands, 2004; Volume 107, pp. 327–331. [Google Scholar] [CrossRef]
  91. Sun, Z.; Deng, Z.-H.; Nie, J.-Y.; Tang, J. RotatE: Knowledge Graph Embedding by Relational Rotation in Complex Space. In Proceedings of the International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  92. Toutanova, K.; Chen, D.; Pantel, P.; Poon, H.; Choudhury, P.; Gamon, M. Representing Text for Joint Embedding of Text and Knowledge Bases. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, Lisbon, Portugal, 17–21 September 2015; Association for Computational Linguistics: Kerrville, TX, USA, 2015; pp. 1499–1509. [Google Scholar]
  93. Hua, M.; Zhang, Y.; Zhang, Q.; Tang, H.; Guo, L.; Lin, Y. KG-IBL: Knowledge Graph Driven Incremental Broad Learning for Few-Shot Specific Emitter Identification. IEEE Trans. Inf. Forensics Secur. 2024, 19, 10016–10028. [Google Scholar] [CrossRef]
  94. Xu, C.; Nayyeri, M.; Alkhoury, F.; Yazdi, H.; Lehmann, J. Temporal Knowledge Graph Completion Based on Time Series Gaussian Embedding. In Proceedings of the Semantic Web–ISWC 2020, Online, 2–6 November 2020; Lecture Notes in Computer Science. Springer: Cham, Switzerland, 2020; Volume 12506. [Google Scholar] [CrossRef]
  95. Consoli, S.; Pezzoli, L.T.; Tosetti, E. Using the GDELT Dataset to Analyse the Italian Sovereign Bond Market. In Proceedings of the Machine Learning, Optimization, and Data Science, LOD 2020, Siena, Italy, 19–23 July 2020; Lecture Notes in Computer Science. Springer: Cham, Switzerland, 2020; Volume 12565. [Google Scholar] [CrossRef]
  96. Zhang, Z.; Cai, J.; Zhang, Y.; Wang, J. Learning Hierarchy-Aware Knowledge Graph Embeddings for Link Prediction. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 3065–3072. [Google Scholar] [CrossRef]
  97. Jiang, C.Y.; Jiang, Y.; Wu, W.Q.; Zheng, Y.T.; Xie, P.J.; Tu, K.W. COMBO: A Complete Benchmark for Open KG Canonicalization. arXiv 2023, arXiv:2302.03905. [Google Scholar]
  98. Lehmann, J.; Isele, R.; Jakob, M.; Jentzsch, A.; Kontokostas, D.; Mendes, P.N.; Hellmann, S.; Morsey, M.; van Kleef, P.; Auer, S.; et al. DBpedia–A large-scale, multilingual knowledge base extracted from Wikipedia. Semant. Web J. 2015, 6, 167–195. [Google Scholar] [CrossRef]
  99. Liu, R.; Wu, L.; Zhang, P. KG-TREAT: Pre-training for Treatment Effect Estimation by Synergizing Patient Data with Knowledge Graphs. In Proceedings of the AAAI Conference on Artificial Intelligence 2024, Vancouver, BC, Canada, 20–27 February 2024; Volume 38, pp. 8805–8814. [Google Scholar] [CrossRef]
  100. Jiang, Z.; Zhong, L.; Sun, M.; Xu, J.; Sun, R.; Cai, H.; Luo, S.; Zhang, Z. Efficient Knowledge Infusion via KG-LLM Alignment. In Proceedings of the Findings of the Association for Computational Linguistics: ACL 2024, Bangkok, Thailand, 11–16 August 2024; Association for Computational Linguistics: Kerrville, TX, USA, 2024; pp. 2986–2999. [Google Scholar]
  101. Wang, Y.; Xie, Q.; Tang, M.; Li, L.; Yuan, J.; Liu, Y. Amazon-KG: A Knowledge Graph Enhanced Cross-Domain Recommendation Dataset. In Proceedings of the 47th International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR ’24), Washington, DC, USA, 14–18 July 2024; Association for Computing Machinery: New York, NY, USA, 2024; pp. 123–130. [Google Scholar] [CrossRef]
  102. Gong, L.; Zhang, H. MR2-KG: A Multi-Relation Multi-Rationale Knowledge Graph for Modeling Software Engineering Knowledge on Stack Overflow. IEEE Trans. Softw. Eng. 2024, 50, 1867–1887. [Google Scholar] [CrossRef]
Figure 1. Process of KG canonicalization.
Figure 1. Process of KG canonicalization.
Mathematics 13 02244 g001
Figure 2. KGs in different scenarios.
Figure 2. KGs in different scenarios.
Mathematics 13 02244 g002
Table 1. Symbols and their specific meanings.
Table 1. Symbols and their specific meanings.
SymbolSpecific Meaning
hHead entity
rRelationship
tTail entity
G Positive sample set
G Negative sample set
γ Interval parameters (margin)
s ( h , r , t ) Score of triplet ( h , r , t )
y i True label
y ^ i Predicted tags
T Test set
rank ( h , r , t ) Ranking of triplet ( h , r , t )
TPRTrue Positive Rate
FPRFalse Positive Rate
Q Query collection
AP ( q ) Average Precision of query q
Element-wise multiplication (Hadamard product)
Re ( · ) Take the real part of a complex number
cosine similarity Cosine similarity
d ( A , B ) Distance between vectors A and B
accuracyAccuracy
MRRMean Reciprocal Rank
Hits@kHit rate of the first k hits
e h , e t Embedding vectors of head and tail entities
r , r rev Embedding vectors of relationship and inverse relationship
Conv2D2D convolution operations
ReLURectified Linear Unit activation function
q h , q r , q t Quaternion representation of head entity, relationship, and tail entity
Quaternion multiplication
JObjective Function (loss function)
C i i th cluster
μ i Center of the i th cluster
D single Single linkage distance
D complete Complete linkage distance
D average Average linkage distance
JObjective function, representing the sum of squared distances from all data points to their respective cluster centers.
nTotal number of data points in the dataset.
KNumber of clusters.
x i i th data point ( i = 1 , 2 , , n ).
μ k Center of the k th cluster (centroid).
r i k Indicator variable indicating whether data point x i belongs to cluster k. If x i belongs to cluster k, r i k = 1 ; otherwise, r i k = 0 .
PProportion of samples predicted as positive by the model that are actually positive.
RProportion of samples that are actually positive classes that are correctly predicted by the model as positive classes.
Table 2. Comparison of different types of knowledge graphs.
Table 2. Comparison of different types of knowledge graphs.
Knowledge Graph TypeMain CharacteristicsApplication ScenariosAdvantagesDisadvantages
Classic Knowledge Graph (CKG)Structured semantic knowledge base based on graphs, storing data in triplet formSocial networks, bioinformatics networks, traffic flow prediction, etc.Rich semantic information, capable of intuitively modeling heterogeneous relationships and supporting interpretable reasoningTraditional methods face limitations in handling complex relationships, long-tail entities, and dynamic knowledge updates
Temporal Knowledge Graph (TKG)Dynamic knowledge graph containing facts that change over timeDrug efficacy analysis, event prediction, etc.Capable of processing temporal information and supporting time-series tasksRelies heavily on the accuracy and clarity of prompts; prompt design and selection are highly demanding
Multi-Modal Knowledge Graph (MMKG)Combines multiple data modalities (e.g., text, images, and videos) to represent factsMedical knowledge graphs (combining text and medical images)Enhances the diversity and accuracy of knowledge representation by integrating multi-modal dataMay produce inaccurate results due to limitations of LLMs (e.g., hallucinations and insufficient attention to modal information)
Inductive Knowledge Graph (IKG)Contains a source KG and a target KG, aiming to transfer structural patterns from the source to the targetRecommendation systems, cross-domain knowledge transferCapable of handling new entities and relations, supporting inductive reasoningMay generate unreliable reasoning results due to LLMs’ hallucinations
Open Knowledge Graph (OKG)Derived from open datasets, user contributions, or internet information, emphasizing openness and scalabilityCyber threat response, mental health observation, geographic relation understanding, etc.Lowers the threshold for knowledge access and enables rapid knowledge expansionData quality is inconsistent, and there are reliability concerns, which may lead to incorrect information generation by LLMs
Domain Knowledge Graph (DKG)Specific to a particular domain, recording domain-specific knowledgeMedicine, electric power, etc.Provides targeted solutions for specific domainsStrong dependency on domain-specific data, with the risk of overfitting
Table 3. Comparison of the 4 categories of KGE methods in the CKG scenario.
Table 3. Comparison of the 4 categories of KGE methods in the CKG scenario.
KGE Method CategoryWorking ProcedureAdvantagesDisadvantages
Methods that use LLMs’ promptsExtract structural and auxiliary information from CKG triplets and convert it into text descriptions to serve as prompts for LLMs. Leverage LLMs’ text generation and reasoning capabilities to complete tasks.Effectively utilize the text generation and reasoning capabilities of LLMs to integrate CKG knowledge embeddings naturally. Useful for complex tasks like knowledge completion and uncovering hidden semantic relationships.Potential information loss or semantic bias during the conversion of triplets to text. Linguistic ambiguity may reduce the accuracy of the original structural information.
Methods for fine-tuning LLMsCollect CKG triplet data, convert them into a format suitable for model training, and fine-tune the LLMs with specific loss functions to optimize performance for the task.Allows the model to learn semantic relationships in CKG triplets, improving performance on CKG tasks and enhancing the model’s professionalism.Requires large amounts of high-quality annotated data. Data collection and preprocessing are time-consuming. Risk of overfitting if data are insufficient or of low quality.
Methods of pre-trained LLMsConvert CKG triplets and background information into text format and use this data for pre-training tasks (e.g., MLM, NSP) to update model parameters.Enhances the model’s understanding of CKG semantics and relationships through pre-training, enriching its knowledge reserves.Conversion to text format may introduce noise or information loss, affecting the accuracy of learned prior knowledge.
Methods to use LLMs as an agentUse LLMs as an agent to interactively explore entities and relations on CKGs and perform reasoning based on retrieved knowledge.Leverages pre-trained knowledge and reasoning patterns in LLMs, combined with CKG structure, for complex knowledge reasoning.Currently less explored and requires further investigation.
Table 4. Research resources related to knowledge graphs based on large language models.
Table 4. Research resources related to knowledge graphs based on large language models.
DatasetsDatasets Resource
ICEWS 05-15 [43]https://dataverse.harvard.edu/dataset.xhtml?persistentId=doi:10.7910/DVN/28075 (accessed on 15 May 2025)
GDELT [44]https://www.gdeltproject.org/data.html (accessed on 15 May 2025)
YAGO [38]https://yago-knowledge.org/ (accessed on 15 May 2025)
UMLS [39]https://www.nlm.nih.gov/research/umls/index.html (accessed on 15 May 2025)
MetaQA [77]https://github.com/yuyuz/MetaQA (accessed on 15 May 2025)
ReVerb45K [78]https://github.com/Gorov/REVERB (accessed on 15 May 2025)
FB15k-237 [60]https://github.com/TimDettmers/ConvE/blob/master/FB15k-237.tar.gz (accessed on 15 May 2025)
WN18RR [70]https://github.com/TimDettmers/ConvE/blob/master/WN18RR.tar.gz (accessed on 15 May 2025)
Table 5. Comparison of relevant metrics for different datasets.
Table 5. Comparison of relevant metrics for different datasets.
Dataset NameData SourceNumber of EntitiesNumber of RelationsNumber of TriplesApplication ScenariosFeatures
ICEWS 05-15Global news media, government reports, sociological research∼50 k-∼400 kTime-series knowledge graphs, event predictionCovers global political, social, and economic events from 2005 to 2015
GDELTGlobal news media reports∼100 k-∼3 millionSocial science research (conflict analysis, political stability)Contains global events, language, and sentiment information
YAGOWikipedia, WordNet, GeoNames∼1.2 million-∼350 kEntity linking, relation extraction, knowledge graph constructionProvides rich semantic information
UMLSNational Library of Medicine (USA)∼300 k-∼1 millionBiomedical text processing, information retrieval, knowledge graph constructionContains a large number of medical terms and concepts
MetaQAMovie database∼4.3 k-∼300 kMachine reading comprehension, question answering system development, multi-hop reasoningIncludes 1-hop, 2-hop, and 3-hop question answering pairs
ReVerb45KWeb page text∼10 k-∼45 kRelation extraction, knowledge graph constructionAutomatically extracted relation instances from web page text
FB15k-237Subset of Freebase knowledge graph15,028237∼310 kKnowledge graph completion, link predictionProcessed subset to reduce overlap between training and testing sets
WN18RRSubset of WordNet knowledge graph40,94311∼93 kKnowledge graph completion, link predictionRevised subset to increase task difficulty
Table 6. Code resources related to knowledge graphs based on large language models.
Table 6. Code resources related to knowledge graphs based on large language models.
ReferenceCode Resource
RecInDial: A Unified Framework for Conversational Recommendation with Pretrained Language Models [43]https://github.com/Lingzhi-WANG/PLM-BasedCRS (accessed on 15 May 2025)
Knowledge Graph Large Language Model (KG-LLM) for Link Prediction [44]https://anonymous.4open.science/r/KG-LLM-FED0 (accessed on 15 May 2025)
THINK-ON-GRAPH: Deep and Responsible Reasoning of Large Language Model on Knowledge Graph [53]https://github.com/IDEA-FinAI/ToG (accessed on 15 May 2025)
Unifying Local and Global Knowledge: Empowering Large Language Models as Political Experts with Knowledge Graphs [38]https://github.com/xymou/PEG (accessed on 15 May 2025)
KC-GenRe: A Knowledge-constrained Generative Re-ranking Method Based on Large Language Models for Knowledge Graph Completion [39]https://github.com/wylResearch/KC-GenRe (accessed on 15 May 2025)
KG-Rank: Enhancing Large Language Models for Medical QA with Knowledge Graphs and Ranking Techniques [77]https://anonymous.4open.science/r/KGQA-270F (accessed on 15 May 2025)
From Large Language Models to Knowledge Graphs for Biomarker Discovery in Cancer [78]https://github.com/rezacsedu/llms_to_kgs_for_cancer (accessed on 15 May 2025)
Multi-perspective Improvement of Knowledge Graph Completion with Large Language Models [60]https://github.com/quqxui/MPIKGC (accessed on 15 May 2025)
UrbanKGent: A Unified Large Language Model Agent Framework for Urban Knowledge Graph Construction [70]https://github.com/usail-hkust/UrbanKGent (accessed on 15 May 2025)
Docs2KG: Unified Knowledge Graph Construction from Heterogeneous Documents Assisted by Large Language Models [63]https://docs2kg.ai4wa.com (accessed on 15 May 2025)
zrLLM: Zero-Shot Relational Learning on Temporal Knowledge Graphs with Large Language Models [55]https://github.com/ZifengDing/zrLLM (accessed on 15 May 2025)
GenTKG: Generative Forecasting on Temporal Knowledge Graph with Large Language Models [57]https://github.com/mayhugotong/GenTKG (accessed on 15 May 2025)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, B.; Fang, Y.; Xu, N.; Hou, S.; Li, X.; Li, Q. Large Language Models for Knowledge Graph Embedding: A Survey. Mathematics 2025, 13, 2244. https://doi.org/10.3390/math13142244

AMA Style

Liu B, Fang Y, Xu N, Hou S, Li X, Li Q. Large Language Models for Knowledge Graph Embedding: A Survey. Mathematics. 2025; 13(14):2244. https://doi.org/10.3390/math13142244

Chicago/Turabian Style

Liu, Bingchen, Yuanyuan Fang, Naixing Xu, Shihao Hou, Xin Li, and Qian Li. 2025. "Large Language Models for Knowledge Graph Embedding: A Survey" Mathematics 13, no. 14: 2244. https://doi.org/10.3390/math13142244

APA Style

Liu, B., Fang, Y., Xu, N., Hou, S., Li, X., & Li, Q. (2025). Large Language Models for Knowledge Graph Embedding: A Survey. Mathematics, 13(14), 2244. https://doi.org/10.3390/math13142244

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop