Next Article in Journal
Supersampling in Render CPOY: Total Annihilation
Previous Article in Journal
Scaling Linearizable Range Queries on Modern Multi-Cores
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

GraphTrace: A Modular Retrieval Framework Combining Knowledge Graphs and Large Language Models for Multi-Hop Question Answering

The Institute of Information Systems, University of Lübeck, 23562 Lübeck, Germany
*
Author to whom correspondence should be addressed.
Computers 2025, 14(9), 382; https://doi.org/10.3390/computers14090382
Submission received: 15 July 2025 / Revised: 2 September 2025 / Accepted: 3 September 2025 / Published: 11 September 2025

Abstract

This paper introduces GraphTrace, a novel retrieval framework that integrates a domain-specific knowledge graph (KG) with a large language model (LLM) to improve information retrieval for complex, multi-hop queries. Built on structured economic data related to the COVID-19 pandemic, GraphTrace adopts a modular architecture comprising entity extraction, path finding, query decomposition, semantic path ranking, and context aggregation, followed by LLM-based answer generation. GraphTrace is compared against baseline retrieval-augmented generation (RAG) and graph-based RAG (GraphRAG) approaches in both retrieval and generation settings. Experimental results show that GraphTrace consistently outperforms the baselines across evaluation metrics, particularly in handling mid-complexity (5–6-hop) queries and achieving top scores in directness during the generation evaluation. These gains are attributed to GraphTrace’s alignment of semantic reasoning with structured KG traversal, combining modular components for more targeted and interpretable retrieval.

1. Introduction

Large language models (LLMs) have demonstrated impressive capabilities in generating fluent, context-aware text across a wide range of natural language processing tasks. However, their effectiveness in knowledge-intensive applications such as question answering remains limited due to their reliance on static, pretrained knowledge. This often results in factual inconsistencies or hallucinations, outputs that deviate from user input, context, or real-world knowledge [1,2,3].
To mitigate this, retrieval-augmented generation (RAG) approaches have been proposed [4,5], which enhance LLMs by incorporating external information retrieval. By integrating external data, RAG reduces factual errors and improves the contextual accuracy [6,7]. Despite these advancements, conventional RAG methods often lack the capacity for structured reasoning or sensemaking, as they typically retrieve from unstructured sources without semantic or relational context [8]. Recent work has explored the use of knowledge graphs (KGs) in RAG systems, referred to as GraphRAG, combining the retrieval power of RAG with the semantic structure of KGs. Approaches such as GraphRAG [9], Think on Graphs (ToG) [10], and LEGO-GraphRAG [11] demonstrate the promise of using graph-based retrieval to better capture relationships between concepts and enable multi-hop reasoning.
Existing GraphRAG methods struggle with complex, multi-hop questions, as they often rely on simple one-hop retrieval strategies such as breadth-first search (BFS), which can miss relevant but distant information or introduce noise [12]. While expanding to larger subgraphs is a potential alternative, it introduces challenges in selecting the appropriate expansion depth and in assessing the relevance of retrieved paths, especially when they are only implicitly related to the input query. Here, GraphTrace is assessed against existing RAG and GraphRAG approaches in both retrieval and generation settings, with a particular focus on the question complexity (measured via hop count) and qualitative answer quality.
This paper presents the architecture of GraphTrace, describes the benchmark and evaluation setup, and discusses empirical results across different task dimensions. The findings show that GraphTrace achieves consistently strong performance, especially in mid-complexity queries, while also revealing insights into the interplay between the graph structure, query formulation, and answer quality. The code is publicly available at https://github.com/AnnaO8/kg-llm-ir (accessed on 9 April 2025).

2. Related Work

Recent years have seen a surge in interest in retrieval-augmented generation (RAG) techniques that combine large language models (LLMs) with external knowledge sources [13]. This section reviews key developments in this area and attempts to situate our approach, GraphTrace, within the landscape of existing methods. For this, Table 1 shows the most important features of existing solutions compared to the features of GraphTrace, the framework that was developed during this work.
Naive RAG represents the simplest retrieval-augmented architecture [4]. A user’s query is forwarded to a dense or sparse retriever (e.g., BM25 or DPR), and a fixed number of top-ranked documents is injected directly into the LLM’s context window. This approach is efficient and easy to implement but suffers from limited reasoning capabilities, especially in complex multi-hop or entity-centric questions, due to the lack of structural awareness and query decomposition. Advanced RAG frameworks aim to improve over this baseline by incorporating entity extraction, context-aware query expansion, or reranking techniques [14]. These pipelines often include NLP preprocessing steps that segment the query into salient components or refine the retrieval intent. However, they remain largely linear in architecture and do not model relationships between documents or entities in a structured manner. Modular RAG systems embrace a compositional architecture where retrieval and reasoning components are cleanly separated and can be dynamically reconfigured [15]. These systems often support plugin-style tool calls, document chaining, and hierarchical retrieval. While modularity improves robustness and interpretability, most implementations operate over unstructured text and do not leverage graph-based representations or topological reasoning.
Microsoft’s GraphRAG introduces the notion of using a knowledge graph (KG) as the retrieval substrate [9]. Queries are mapped to entities or relations in the KG, and relevant subgraphs (e.g., k-hop neighborhoods or semantic paths) are extracted and linearized into text for prompt injection. While this method benefits from structural compression and thus improved relevance and factual grounding, it remains inherently static: communities are precomputed, and the summarization step flattens fine-grained relationships between nodes. Consequently, the system lacks support for query-specific exploration and interpretable path reasoning. LazyGraphRAG addresses some of these limitations by introducing adaptive graph traversal and cost-aware reasoning [8]. The system incrementally explores the graph and performs retrieval steps only as needed, balancing latency and answer quality. It combines best-first search based on vector similarity with breadth-first exploration along graph neighborhoods. While this reduces the computational overhead, it still relies heavily on vector-based heuristics to guide the search process. The absence of symbolic reasoning and explicit path modeling means that relevance is approximated rather than semantically evaluated. Moreover, neither GraphRAG nor LazyGraphRAG supports the decomposition of the input query into smaller subquestions or their alignment with specific paths in the graph. As a result, both approaches struggle with complex, multifaceted queries that require the assembly of information from multiple, semantically distinct subgraphs.
Taken together, the approaches each offer valuable contributions: Naive RAG excels in simplicity and speed, Advanced RAG improves the input quality through preprocessing, Modular RAG emphasizes composability, GraphRAG enables structured retrieval, and LazyGraphRAG introduces adaptivity. Another recent system, graph retrieval-augmented generation (GRAG), proposes a dual-view approach that combines soft prompting and symbolic exploration over KGs [16]. GRAG constructs an ego-centric subgraph around query entities using relevance-guided expansion and then transforms these into two views: a hard prompt representing textualized triples and a soft prompt derived from learned graph embeddings. These are jointly injected into the LLM for answer generation. While GRAG introduces a flexible prompting mechanism and leverages both symbolic and neural representations, it does not explicitly decompose queries or semantically align subgraphs with query components. Moreover, the ego-graph construction lacks robust path modeling and canonicalization, making it challenging to assess whether the retrieved context truly matches the user’s information needs—particularly for multifaceted queries that span divergent or converging subgraphs.
LEGO-GraphRAG extends the modular RAG paradigm to graph-based retrieval by introducing a compositional framework that supports plug-and-play components for entity recognition, graph traversal, semantic ranking, and context integration [11]. Each module in LEGO-GraphRAG can be configured independently, enabling flexible adaptation to task-specific requirements or resource constraints. Notably, it supports the semantic reranking of paths and offers basic graph traversal strategies beyond simple BFS. However, the system relies primarily on predefined modules, without a unified semantic alignment mechanism across subqueries and retrieved paths.
Despite all of these advancements, a fundamental issue in RAG lies in assessing the relevance of multi-hop paths [6,17,18]. These paths may not exhibit a direct or obvious connection to the query but can nevertheless contain crucial contextual information. Evaluating such latent relevance is difficult: simple heuristics based on path lengths, shared nodes, or edge frequencies often fail to capture the semantic alignment between query intent and path content. What is still lacking is a system that combines explicit entity extraction, semantic path finding and ranking, adaptive context construction, and LLM-compatible output formatting into a cohesive, end-to-end pipeline. This fragmentation motivates the development of an integrated framework that draws from the strengths of the existing approaches while addressing their individual limitations. By merging structured retrieval, decomposition strategies, and semantic ranking with LLM-based generation, our GraphTrace system can support complex, multi-hop queries more effectively than any of its components alone.
This work builds upon a knowledge graph about the impacts of the COVID-19 pandemic [19] on the economy, introduced by [20], which was constructed from a curated set of 62 reports and economic bulletins by organizations such as the World Bank, IMF, and European Central Bank. The data were structured using LLM-generated question–answer pairs to facilitate triple extraction and entity linking via Wikidata. The resulting KG comprises over 30,000 triples and serves as the foundation for the benchmark dataset used in our evaluation. The question–answer dataset, derived from this KG, is explained further in Section 4.

3. Methodology Overview

This study adopts an experimental computer science methodology. We present GraphTrace as a research artifact that is developed as a retrieval framework for multi-hop question answering. Following this approach, we (i) introduce the design of the artifact, (ii) demonstrate its use on a domain-specific KG, and (iii) assess its performance through standardized retrieval and generation evaluation.
Existing GraphRAG methods face significant challenges in handling complex, multi-hop questions. Most approaches rely on one-hop breadth-first search (BFS) to retrieve information from the immediate neighborhood of extracted entities [12]. This strategy is often inadequate for two primary reasons: (1) the direct neighborhood may contain either insufficient information or excessive noise if not properly filtered, and (2) many real-world questions require multi-hop reasoning to access semantically related but more distant nodes in the KG. While retrieving larger subgraphs is a potential alternative, it presents its own challenges, particularly in determining the optimal depth and direction of graph expansion. Furthermore, assessing the relevance of multi-hop paths is inherently difficult, as these paths may not have an explicit connection to the input query, yet they can contain information that is essential in generating an accurate answer.
To address these limitations, we propose GraphTrace, a modular retrieval pipeline designed to enhance semantic understanding and reasoning over KGs. The framework consists of five core components: entity extraction, path finding, query decomposition, semantic path ranking, and context aggregation. Together, these components enable the semantically guided exploration of the KG and the targeted integration of relevant information, tailored to the input query. Figure 1 provides an overview of the conceptual design of the GraphTrace pipeline, highlighting the individual stages and their interactions. To illustrate how this design operates in practice, Figure 2 presents an example.

3.1. Entity Extraction

The starting point of the GraphTrace system is the user query, which is processed by the first component: entity extraction. The goal of this component is to identify key entities in the input query and determine appropriate starting points in the KG. The input is the user query in unstructured text form, and the output is a list of entities represented as strings. This step is performed using the LLM provided by OpenAI (gpt-4o-mini) [21]. The model is prompted with a system message containing detailed instructions for the entity extraction task, including the expected output format, a JSON object consisting of a list of extracted entity strings. Figure 3 illustrates the entity extraction process, including a concrete example from the utilized question–answer dataset.

3.2. Path Finding

The extracted entities serve as input to the next component: path finding. As illustrated in Figure 4, this component first verifies whether the extracted entities exist as nodes in the KG. This verification is performed by executing Cypher queries that match nodes based on their name attributes. For entities found in the KG, the next step is to identify all paths between them using the allShortestPaths((e1)-[*]-(e2)) function. This retrieves all shortest paths between entity pairs, regardless of the relationship direction or path length. To maintain efficiency and reduce noise, self-loops (i.e., paths where the start and end nodes are the same) are explicitly excluded. The use of allShortestPaths() provides a balance between computational efficiency and comprehensive coverage, capturing diverse yet concise semantic connections between verified entities. Following path extraction, a canonicalization step is performed to eliminate redundant representations and ensure consistency in the retrieved paths. The output of this component is a list of extracted paths (Path 1, …, Path n), which are subsequently passed to downstream components. An example of the process is depicted in Figure 4.

3.3. Query Decomposition

In addition to entity extraction, the user query is also passed to the query decomposition component. The purpose of this step is to divide the original query into smaller, manageable subqueries, thereby reducing its overall complexity and facilitating more precise retrieval. The input to this component is the user query in unstructured text form, and the output is a list of subqueries represented as strings. The model is prompted with a system message containing detailed instructions for query decomposition, including the expected output format—a JSON object consisting of a list of strings corresponding to the generated subqueries. This structured output format ensures consistency and simplifies integration with subsequent pipeline components. The user query is then provided as a message to the LLM, triggering the decomposition process. Figure 5 illustrates the query decomposition component, including an example derived from the utilized question–answer dataset.

3.4. Path Ranking

The subsequent component in the GraphTRACE pipeline is path ranking, which takes as input the previously extracted paths and the generated subqueries. Its primary objective is to evaluate and rank all paths for each subquery based on semantic relevance. Scoring and ranking are performed using the ms-marco-MiniLM-L-6-v2 CrossEncoder model [22]. This model is specifically trained for task ranking and computes a relevance score for each (subquery, path) pair by jointly encoding both elements as a single input. This joint encoding allows the model to capture subtle semantic dependencies and contextual relationships, making it well suited for comparing a fixed subquery against multiple candidate paths. The MiniLM variant offers a favorable balance between computational efficiency and ranking quality, which is why it was selected for this task. For each subquery, the model scores all associated paths and ranks them in descending order of predicted relevance. The output of this component is a ranked list of paths for each subquery, from most to least relevant. This process is illustrated in Figure 6, including an example visualization of the scoring and ranking procedure.

3.5. Aggregation

The final component of the retrieval pipeline is the aggregation module. In this step, the top three most relevant paths for each subquery, as determined by the path ranking component, are selected and aggregated to form the final retrieved context. Since the subqueries represent logical subdivisions of the original query, aggregating the most relevant paths for each subquery ensures that the resulting context is comprehensive and semantically aligned with the original information need. This aggregated set of paths serves as the foundation for subsequent answer generation in the overall GraphRAG pipeline. A complete example of this aggregation process is illustrated in Figure 7.

4. Evaluation

The following section presents the evaluation of the proposed retrieval framework, GraphTrace, in comparison to the baseline RAG and GraphRAG methods. The evaluation is conducted using an early version of the EcoRAG dataset [23], captured during its initial development phase before the full set of question–answer (QA) pairs was available. It is a multi-hop economic QA dataset constructed over the Economic_KG, the domain-specific knowledge graph. EcoRAG is designed for deep, multi-hop retrieval with queries requiring up to seven hops and includes complex subgraph topologies such as converging, divergent, and linear patterns.

4.1. Evaluation Setup

In this section, the setup for the evaluation is explained in detail. First, the dataset used for the evaluation of the system is described. After this, the different approaches’ methods of accessing the KG, acquiring information from it, and generating answers are discussed.
Question–Answer Dataset: The question–answer dataset created by [20] includes structured queries and annotated ground-truth paths in the graph. It consists of three types of graph reasoning structures—linear, converging, and divergent—each reflecting increasing levels of complexity in semantic graph traversal.
  • Linear Paths (Chain Reasoning)
    Linear paths represent a direct sequence of connected entities and relationships. Each step in the path depends on the preceding one, forming a straight-line inference structure.
    Structure: A → B → C → D
  • Converging Paths (Directed Acyclic Graphs—DAGs)
    Converging paths involve multiple reasoning branches that lead to a common node. These structures are used when synthesizing multiple sources of information to reach a unified conclusion.
    Structure: (A → B → C), (D → E → C)
  • Divergent Paths (Polytrees)
    Divergent paths originate from a single entity that connects to multiple downstream branches, each representing an independent line of inference.
    Structure: A → B, A → C, A → D
In contrast, the majority of existing GraphRAG approaches [9] follow a multi-stage pipeline that typically begins with a corpus of text documents and first constructs a KG through an initial KG construction step. This graph is then used for downstream information retrieval and answer generation.
Since the KG construction step has already been performed and documented by [20], this pipeline directly adopts their resulting KG as the dataset. This approach enables a focused evaluation of the retrieval and generation components without reintroducing variability from the construction phase. As part of the evaluation methodology, the following methods are assessed. Each approach differs in how it accesses the KG, performs retrieval, and generates answers.
  • Naive RAG: This is a basic retrieval-augmented generation approach that performs a semantic search over a CSV-based knowledge graph. It uses SentenceTransformer (all-MiniLM-L6-v2) with FAISS [24] to embed and index KG triples. For each query, the top six most relevant triples are retrieved and combined with the query and then passed to OpenAI’s gpt-4o-mini to generate an answer.
  • Naive RAG with Subquery: This extends the basic Naive RAG approach by decomposing complex queries into simpler subqueries using an LLM (gpt-4o-mini). Each subquery is processed individually to improve the retrieval effectiveness through query simplification.
  • Hybrid RAG: This combines dense (vector-based) and sparse (BM25) retrieval. Results are merged and used as context for generation.
  • Rerank RAG: This retrieves top triples using dense embeddings and then reranks them with a CrossEncoder to improve the result quality before answer generation.
  • Naive GraphRAG: This applies BFS and DFS over the KG starting from entities extracted via KeyBERT [25]. Aggregated paths are used for generation.
  • KG RAG: Based on [26], this is a graph-based retrieval method that uses the LLM gpt-4o-mini to guide the step-by-step exploration of a knowledge graph. Based on the input query, the LLM creates a plan to decide whether to explore nodes or relationships. For node exploration, it finds the top five candidates using vector search, and the LLM selects the most relevant ones. For relationships, it identifies paths between important nodes and verifies their relevance. This process continues until enough information is gathered to answer the query. If the plan fails after three revisions, the system stops and does not return an answer.
  • Think on Graphs (ToG): This is graph-based retrieval method from [10] that explores the knowledge graph iteratively. Since the original version was designed for Freebase [27] or Wikidata [28], this study adapts it to work with the Economic_KG by integrating KeyBERT for entity recognition. The system starts from identified key entities and uses beam search to explore surrounding nodes and relations. Irrelevant results are filtered through pruning, and reasoning steps (guided by gpt-4o-mini) determine whether enough context has been retrieved or more exploration is needed. This process continues until the system decides that an answer can be generated.
Given that the proposed approach comprises two main components—retrieval and generation—it is essential to evaluate these stages independently. This separation enables a more fine-grained analysis of system performance, allowing us to identify the strengths and limitations of each component in contributing to the overall task of knowledge-augmented question answering.

4.2. Retrieval Evaluation

The retrieval performance is evaluated using the metrics of the mean reciprocal rank (MRR), mean average precision (MAP), and Hit@10. These metrics assess the quality of the retrieved information across different methods. The MRR is defined as follows:
MRR = 1 | Q | i = 1 | Q | 1 rank i
where rank i refers to the rank position of the first relevant document for the i-th query.
Each query q Q is assigned a reciprocal rank score between 0 and 1, reflecting the rank position of the first relevant document in the result list. The MRR aggregates these scores into a single value within the interval [0, 1] that represents the average ranking performance across all queries in the dataset. The higher the MRR, the better the system’s performance in ranking relevant documents at the top. However, it is important to note that the MRR considers only the first relevant result for each query.
The MAP is defined as follows:
MAP = q = 1 Q AP ( q ) Q
where Q is the number of queries .
The MAP evaluates both the rank positions and the number of relevant documents in the result list across the entire set of queries. For this, it computes the AP for each query by identifying all positions of relevant documents and calculating their precision at each of these positions. Based on this, the average of these precision values yields the AP for each query. For the MAP, the mean value of all AP scores is derived and lies in the interval [0, 1]. As with the MRR, a higher value here also indicates better system performance. In contrast to the MRR, the MAP includes all relevant results of all queries.
The Hit@10 indicates, for each query, whether at least one relevant document appears in the top 10 results. This metric uses an indicator function to check whether the condition is true or not and returns a respective value of 1 or 0. The overall Hit@10 score is calculated as the mean value over all queries in the dataset and lies in the interval [0, 1]. The Hit@10 does not take the exact positions of relevant documents into account, but it is useful in evaluating whether the system is capable of retrieving relevant information in the first 10 positions. The same applies here as for the MRR and MAP: a higher value indicates better performance.
Table 2 presents the results for the aggregated QA dataset, which combines the converging, divergent, and linear subsets, resulting in a total of 267 question–answer pairs. GraphTRACE achieves the best performance in terms of the MRR (0.4477) and MAP (0.1906) and remains competitive for the Hit@10 (0.8127). To confirm the robustness of the results, the experiments were repeated five times. The performance remained stable across runs, with standard deviations below 0.015 across all metrics (MRR = 0.4429 ± 0.0102, MAP = 0.1915 ± 0.0014, Hit@10 = 0.8230 ± 0.0112). Naive RAG with Subquery shows notable improvements over Naive RAG, especially in the MAP (0.1136 vs. 0.0873), highlighting the benefit of query decomposition. Hybrid RAG performs similarly to Naive RAG, with only minor differences.
Table 3 presents the retrieval evaluation results for the converging dataset, which contains 90 question–answer pairs. GraphTRACE achieves the best performance in both the MRR ( 0.484 ) and MAP ( 0.2314 ) and performs competitively regarding the Hit@10 ( 0.8444 ). Naive RAG + Subquery outperforms standard Naive RAG across all metrics and achieves the highest Hit@10 score among all evaluated methods ( 0.8778 ), demonstrating the benefit of query decomposition. Hybrid RAG performs similarly to Naive RAG, with only minor differences across metrics. Rerank RAG shows strong performance regarding the Hit@10 ( 0.8333 ), but its MRR score ( 0.3102 ) is lower than that of the other RAG-based baselines. Naive GraphRAG and KG RAG exhibit poor performance across all retrieval metrics.
Table 4 presents the retrieval evaluation results for the divergent dataset, which consists of 90 question–answer pairs. GraphTRACE shows competitive performance in the MRR ( 0.4149 ) and Hit@10 ( 0.8000 ), while achieving the highest score for the MAP ( 0.1750 ). Naive RAG attains the highest MRR ( 0.4261 ) but is outperformed by Naive RAG + Subquery in both the MAP ( 0.1164 vs. 0.1038 ) and Hit@10 ( 0.8333 vs. 0.8000 ), highlighting the benefit of query decomposition. Hybrid RAG surpasses Naive RAG with Subquery in terms of the MRR but performs worse in the other metrics. Rerank RAG produces similar results to Hybrid RAG for both the MRR and MAP, while achieving the best Hit@10 score ( 0.8444 ). Once again, the graph-based baselines—Naive GraphRAG and KG RAG—show poor performance across all retrieval metrics.
Table 5 presents the retrieval evaluation results for the linear dataset, which consists of 87 question–answer pairs. GraphTRACE achieves the highest scores in both the MRR ( 0.4442 ) and MAP ( 0.1644 ). For the Hit@10, it ties with Naive RAG ( 0.7931 ) and ranks third overall. Naive RAG shows solid overall performance and slightly outperforms Naive RAG + Subquery in the MRR ( 0.3464 vs. 0.3336 ) and Hit@10 ( 0.7931 vs. 0.7471 ). Hybrid RAG performs similarly to Naive RAG, with marginal improvements in the MAP and Hit@10. Rerank RAG achieves the best Hit@10 score ( 0.8506 ) but ranks lower than Naive RAG + Subquery in the MRR and MAP. As in previous evaluations, the graph-based methods—Naive GraphRAG and KG RAG—perform significantly worse across all metrics.
Following the dataset-level analysis, this section evaluates the impact of the hop count on the retrieval performance. The hop count refers to the number of relationships traversed in the KG to answer a query, with higher counts typically indicating greater complexity. Each dataset (converging, divergent, linear) includes questions with varying hop counts. Figure 8 shows the distribution, and the retrieval metrics (MRR, MAP, Hit@10) are analyzed per hop level. Table 6, Table 7 and Table 8 show the retrieval results across different hop counts for the converging, divergent, and linear datasets. Each is evaluated using the MRR, MAP, and Hit@10. Figure 8 shows the distribution of the hop counts, while Figure 9 visualizes the performance trends by hop level.
Figure 8 also summarizes the hop-based complexity across datasets. The converging set is dominated by seven-hop questions (47), followed by six (23) and eight (17). The divergent set is mostly six-hop (70), with a few five-hop (10) and seven-hop (8) questions. In the linear set, six-hop questions prevail (77), with very few five-hop (6) and four-hop (4) instances. In the converging dataset, GraphTRACE performs best in terms of the MRR and MAP for five- to seven-hop questions—the most frequent categories—while Naive RAG + Subquery achieves the top Hit@10 scores for seven- and eight-hop questions. Most other methods show a drop in performance beyond six hops. Graph-based methods (Naive GraphRAG, KG RAG) perform poorly across all metrics. In the divergent dataset, dominated by six-hop questions, Naive RAG performs well in the MRR, while GraphTRACE achieves the highest MAP across nearly all hop levels. Rerank RAG delivers the strongest Hit@10 performance, particularly at higher hop counts, suggesting that reranking is useful in retrieving dispersed or weakly connected information. The graph-based baselines again show the weakest results. In the linear dataset, where the hop counts range from four to six, GraphTRACE consistently outperforms other methods in both the MRR and MAP. Although Rerank RAG achieves the highest Hit@10 at six hops, GraphTRACE performs strongly overall. The simpler structure of this dataset results in smaller performance gaps among methods.
Overall, GraphTRACE performs robustly across datasets and hop levels, particularly on medium-complexity queries (5–7 hops). Naive RAG + Subquery is effective in converging topologies, while Rerank RAG excels in scenarios with less direct connectivity. In contrast, Naive GraphRAG and KG RAG show consistently poor performance, regardless of the hop count or dataset.

4.3. Generation Evaluation

To evaluate the generated answers, LLMs are employed as evaluators. Prior studies [29,30] have demonstrated the effectiveness of LLMs in assessing natural language generation, often achieving state-of-the-art or competitive results when compared to human judgment [9]. In this evaluation, retrieval metrics provide quantitative comparisons based on ground truth data, whereas generation metrics focus on qualitative aspects. For the latter, an LLM is used as a judge to perform head-to-head comparisons between the generated answers. The selected evaluation criteria include comprehensiveness, diversity, empowerment, and directness, as proposed by Edge et al. [9]. These metrics collectively aim to capture the sensemaking quality of the answers.
  • Comprehensiveness measures the level of detail and completeness in the answer. A comprehensive response is consistent, covers all relevant aspects of the query, and incorporates contextual and adjacent information aligned with the question’s complexity.
  • Diversity evaluates the variety and richness of content. This includes the presentation of different perspectives, insights, or arguments and ensures that the answer offers novel information rather than paraphrasing existing responses.
  • Empowerment assesses the extent to which an answer enables the reader to understand the topic and make informed decisions. This metric emphasizes critical thinking, user autonomy, and the capacity to promote reflective understanding.
  • Directness captures how precisely the answer addresses the core of the question. A direct response avoids digressions and unnecessary elaboration, while remaining clear, structured, and specific. Excessive detail is only acceptable if it supports comprehension.
Since directness can conflict with comprehensiveness and diversity, it is not expected that any one method will outperform others across all four criteria [9]. Therefore, an additional overall score is computed to determine a final winner in the comparative evaluations. For the evaluation process, the LLM is provided with the original question, a description of the target evaluation metric, and the answers generated by each competing approach. The LLM is then prompted to assess which answer best satisfies the specified metric, accompanied by a brief explanation for its choice. Following the protocol proposed by [9], the LLM is instructed to declare a single winner if and only if a clear distinction exists. If the responses are fundamentally similar and the differences are negligible, the evaluation results in a tie.
To mitigate the inherent stochasticity of LLM outputs, the evaluation is repeated three times per instance, and the final scores are computed by averaging the outcomes across runs. The generation evaluation assesses the quality of answers produced by each method using four qualitative metrics: comprehensiveness, diversity, empowerment, and directness. Evaluations were conducted using gpt-4o-mini, which was tasked with identifying the best answer per criterion. To ensure robustness, each evaluation was repeated three times, and the average relative win rates were calculated (Table 9).
Across all datasets, a different method dominates each criterion. Naive RAG with Subqueryachieves the highest win rates for diversity in all three datasets, likely due to its query decomposition strategy, which encourages diverse subquery paths. GraphTRACE is the consistent top performer for directness, suggesting that its path reranking mechanism helps to generate more focused and concise answers. Naive GraphRAG performs best in comprehensiveness for two datasets and ranks highest overall in the converging set. Meanwhile, KG RAG produces the most empowering answers in two datasets.
Interestingly, the graph-based methods Naive GraphRAG and KG RAG, despite underperforming in the retrieval metrics, excel in generation quality. This contrast may be explained by their structured, exploration-based retrieval, which, when successful, enables the generation of more comprehensive and informative answers. KG RAG, for example, follows a strict exploration plan and returns no answer if the plan fails—degrading the retrieval metrics but benefiting generation quality when the plan succeeds.
In contrast, methods like Naive RAG, Hybrid RAG, and Rerank RAG perform moderately or poorly across most generation metrics. GraphTRACE stands out as the most balanced approach, consistently ranking among the top two for most metrics and demonstrating both precision (directness) and general quality across datasets.
To complement the LLM-based evaluation, we conducted a small-scale human study with 40 samples (stratified by dataset and hop count) rated by two independent annotators. The annotators used the same criteria as in our LLM evaluation. We observed an inter-annotator agreement rate of 67%. The human evaluation confirms the LLM-based evaluation: graph-based approaches achieve strong results in answer generation. Notably, GraphTrace was rated even more highly in the “empowerment” criterion by the human annotators than in the LLM-based assessment.

5. Conclusions

This paper introduces GraphTrace, a modular retrieval framework that combines the structured reasoning capabilities of knowledge graphs (KGs) with the generative power of large language models (LLMs) to answer complex, multi-hop questions. By leveraging a domain-specific economic KG centered on the impacts of the COVID-19 pandemic, GraphTrace was designed to navigate and extract relevant information across multiple graph hops and entity relationships. The framework integrates key components—entity extraction, query decomposition, path ranking, and aggregation—into a cohesive pipeline that enables precise retrieval and informed answer generation. Across a range of evaluation metrics and datasets, GraphTrace consistently outperformed the baseline RAG and GraphRAG methods, particularly excelling in medium-complexity questions (5–6 hops) and in generating direct, focused answers.
Despite these strengths, challenges remain. GraphTrace shows diminished performance on low- and high-complexity questions, and generation metrics reveal room for improvement in areas such as empowerment and comprehensiveness. These findings underscore the importance of continued refinement in both evaluation design and system adaptability. Future work will focus on extending the benchmark to include more diverse question types and hop distributions, as well as incorporating adaptive mechanisms into the retrieval pipeline. Furthermore, the evaluation will be extended by considering additional aspects such as correctness, hallucination rates, and semantic similarity to provide an even more comprehensive assessment of the framework. These enhancements will aim to further improve the system’s robustness and effectiveness in real-world knowledge retrieval scenarios.

Author Contributions

Conceptualization, A.O. and H.K.; methodology, A.O., H.K., J.G. and S.G.; software, A.O.; validation, A.O., H.K. and J.G.; formal analysis, A.O.; investigation, A.O., H.K., A.-L.K. and J.G.; resources, A.O., H.K. and A.-L.K.; data curation, A.O. and H.K.; writing—original draft preparation, A.O., H.K. and A.-L.K.; writing—review and editing, A.O., H.K. and A.-L.K.; visualization, A.O.; supervision, H.K., S.G. and J.G.; project administration, S.G. and J.G. funding acquisition, S.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the German Research Foundation under project number 490998901.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhang, Y.; Li, Y.; Cui, L.; Cai, D.; Liu, L.; Fu, T.; Huang, X.; Zhao, E.; Zhang, Y.; Chen, Y.; et al. Siren’s song in the AI ocean: A survey on hallucination in large language models. arXiv 2023, arXiv:2309.01219. [Google Scholar] [CrossRef]
  2. Adlakha, V.; BehnamGhader, P.; Lu, X.H.; Meade, N.; Reddy, S. Evaluating correctness and faithfulness of instruction-following models for question answering. Trans. Assoc. Comput. Linguist. 2024, 12, 681–699. [Google Scholar] [CrossRef]
  3. Liu, T.; Zhang, Y.; Brockett, C.; Mao, Y.; Sui, Z.; Chen, W.; Dolan, B. A token-level reference-free hallucination detection benchmark for free-form text generation. arXiv 2021, arXiv:2104.08704. [Google Scholar]
  4. Lewis, P.; Perez, E.; Piktus, A.; Petroni, F.; Karpukhin, V.; Goyal, N.; Küttler, H.; Lewis, M.; Yih, W.t.; Rocktäschel, T.; et al. Retrieval-augmented generation for knowledge-intensive nlp tasks. Adv. Neural Inf. Process. Syst. 2020, 33, 9459–9474. [Google Scholar]
  5. Gao, Y.; Xiong, Y.; Gao, X.; Jia, K.; Pan, J.; Bi, Y.; Dai, Y.; Sun, J.; Wang, H.; Wang, H. Retrieval-augmented generation for large language models: A survey. arXiv 2023, arXiv:2312.10997. [Google Scholar]
  6. Tang, Y.; Yang, Y. Multihop-rag: Benchmarking retrieval-augmented generation for multi-hop queries. arXiv 2024, arXiv:2401.15391. [Google Scholar]
  7. Peng, B.; Zhu, Y.; Liu, Y.; Bo, X.; Shi, H.; Hong, C.; Zhang, Y.; Tang, S. Graph retrieval-augmented generation: A survey. arXiv 2024, arXiv:2408.08921. [Google Scholar] [CrossRef]
  8. Edge, D.; Trinh, H.; Larson, J. LazyGraphRAG: Setting a New Standard for Quality and Cost. 2024. Available online: https://www.microsoft.com/en-us/research/blog/lazygraphrag-setting-a-new-standard-for-quality-and-cost/ (accessed on 1 July 2025).
  9. Edge, D.; Trinh, H.; Cheng, N.; Bradley, J.; Chao, A.; Mody, A.; Truitt, S.; Metropolitansky, D.; Ness, R.O.; Larson, J. From Local to Global: A Graph RAG Approach to Query-Focused Summarization. arXiv 2025, arXiv:2404.16130. [Google Scholar] [CrossRef]
  10. Sun, J.; Xu, C.; Tang, L.; Wang, S.; Lin, C.; Gong, Y.; Ni, L.M.; Shum, H.Y.; Guo, J. Think-on-graph: Deep and responsible reasoning of large language model on knowledge graph. arXiv 2023, arXiv:2307.07697. [Google Scholar]
  11. Cao, Y.; Gao, Z.; Li, Z.; Xie, X.; Zhou, K.; Xu, J. LEGO-GraphRAG: Modularizing Graph-based Retrieval-Augmented Generation for Design Space Exploration. arXiv 2025, arXiv:2411.05844. [Google Scholar] [CrossRef]
  12. Liu, R.; Jiang, H.; Yan, X.; Tang, B.; Li, J. PolyG: Effective and Efficient GraphRAG with Adaptive Graph Traversal. arXiv 2025, arXiv:2504.02112. [Google Scholar]
  13. Ouyang, L.; Wu, J.; Jiang, X.; Almeida, D.; Wainwright, C.; Mishkin, P.; Zhang, C.; Agarwal, S.; Slama, K.; Ray, A.; et al. Training language models to follow instructions with human feedback. Adv. Neural Inf. Process. Syst. 2022, 35, 27730–27744. [Google Scholar]
  14. Yu, H.; Gan, A.; Zhang, K.; Tong, S.; Liu, Q.; Liu, Z. Evaluation of Retrieval-Augmented Generation: A Survey. In Proceedings of the Big Data; Zhu, W., Xiong, H., Cheng, X., Cui, L., Dou, Z., Dong, J., Pang, S., Wang, L., Kong, L., Chen, Z., Eds.; Springer: Singapore, 2025; pp. 102–120. [Google Scholar]
  15. Gao, Y.; Xiong, Y.; Wang, M.; Wang, H. Modular RAG: Transforming RAG Systems into LEGO-like Reconfigurable Frameworks. arXiv 2024, arXiv:2407.21059. [Google Scholar] [CrossRef]
  16. Hu, Y.; Lei, Z.; Zhang, Z.; Pan, B.; Ling, C.; Zhao, L. GRAG: Graph Retrieval-Augmented Generation. arXiv 2024, arXiv:2405.16506. [Google Scholar] [CrossRef]
  17. Shi, Y.; Tan, Q.; Wu, X.; Zhong, S.; Zhou, K.; Liu, N. Retrieval-enhanced Knowledge Editing in Language Models for Multi-Hop Question Answering. In Proceedings of the 33rd ACM International Conference on Information and Knowledge Management, New York, NY, USA, 21–25 October 2024; CIKM ’24. pp. 2056–2066. [Google Scholar] [CrossRef]
  18. Mavi, V.; Jangra, A.; Jatowt, A. Multi-hop Question Answering. Found. Trends Inf. Retr. 2024, 17, 457–586. [Google Scholar] [CrossRef]
  19. Gruenwald, L.; Jain, S.; Groppe, S. (Eds.) Leveraging Artificial Intelligence in Global Epidemics; Elsevier: Amsterdam, The Netherlands, 2021. [Google Scholar]
  20. Khorashadizadeh, H.; Mihindukulasooriya, N.; Ranji, N.; Ezzabady, M.; Ieng, F.; Groppe, J.; Benamara, F.; Groppe, S. Construction and Canonicalization of Economic Knowledge Graphs with LLMs. In Proceedings of the International Knowledge Graph and Semantic Web Conference; Springer: Cham, Switzerland, 2024; pp. 334–343. [Google Scholar]
  21. Hurst, A.; Lerer, A.; Goucher, A.P.; Perelman, A.; Ramesh, A.; Clark, A.; Ostrow, A.; Welihinda, A.; Hayes, A.; Radford, A.; et al. Gpt-4o system card. arXiv 2024, arXiv:2410.21276. [Google Scholar] [CrossRef]
  22. Sentence Transformers Team msmarco-MiniLM-L6-en-de-v1. 2021. Available online: https://huggingface.co/cross-encoder/msmarco-MiniLM-L6-en-de-v1 (accessed on 30 June 2025).
  23. Khorashadizadeh, H.; Tiwari, S.; Benamara, F.; Mihindukulasooriya, N.; Groppe, J.; Sahri, S.; Ezzabady, M.; Ieng, F.; Groppe, S. EcoRAG: A Multi-hop Economic QA Benchmark for Retrieval Augmented Generation Using Knowledge Graphs. In Proceedings of the Natural Language Processing and Information Systems (NLDB), Kanazawa, Japan, 4–6 July 2025. [Google Scholar]
  24. Cross-Encoder Team all-MiniLM-L6-v2. 2021. Available online: https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2 (accessed on 30 June 2025).
  25. Grootendorst, M. KeyBERT: Minimal Keyword Extraction with BERT. 2020. Available online: https://www.maartengrootendorst.com/blog/keybert/ (accessed on 1 July 2025).
  26. Sanmartin, D. Kg-rag: Bridging the gap between knowledge and creativity. arXiv 2024, arXiv:2405.12035. [Google Scholar] [CrossRef]
  27. Bollacker, K.; Evans, C.; Paritosh, P.; Sturge, T.; Taylor, J. Freebase: A collaboratively created graph database for structuring human knowledge. In Proceedings of the 2008 ACM SIGMOD International Conference on Management of Data, Vancouver, BC, Canada, 9–12 June 2008; pp. 1247–1250. [Google Scholar]
  28. Vrandečić, D.; Krötzsch, M. Wikidata: A free collaborative knowledgebase. Commun. ACM 2014, 57, 78–85. [Google Scholar] [CrossRef]
  29. Wang, J.; Liang, Y.; Meng, F.; Sun, Z.; Shi, H.; Li, Z.; Xu, J.; Qu, J.; Zhou, J. Is chatgpt a good nlg evaluator? a preliminary study. arXiv 2023, arXiv:2303.04048. [Google Scholar] [CrossRef]
  30. Zheng, L.; Chiang, W.L.; Sheng, Y.; Zhuang, S.; Wu, Z.; Zhuang, Y.; Lin, Z.; Li, Z.; Li, D.; Xing, E.; et al. Judging llm-as-a-judge with mt-bench and chatbot arena. Adv. Neural Inf. Process. Syst. 2023, 36, 46595–46623. [Google Scholar]
Figure 1. Conceptual architecture of the GraphTrace retrieval pipeline. The framework consists of five modular components (1. Entity Extraction, 2. Path Finding, 3. Query Decomposition, 4. Path Ranking, 5. Aggregation).
Figure 1. Conceptual architecture of the GraphTrace retrieval pipeline. The framework consists of five modular components (1. Entity Extraction, 2. Path Finding, 3. Query Decomposition, 4. Path Ranking, 5. Aggregation).
Computers 14 00382 g001
Figure 2. Step-by-step example to illustrate GraphTrace’s conceptual design.
Figure 2. Step-by-step example to illustrate GraphTrace’s conceptual design.
Computers 14 00382 g002
Figure 3. Entity extraction component.
Figure 3. Entity extraction component.
Computers 14 00382 g003
Figure 4. Path finding component.
Figure 4. Path finding component.
Computers 14 00382 g004
Figure 5. Query decomposition component; see https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2 (accessed on 10 July 2025).
Figure 5. Query decomposition component; see https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2 (accessed on 10 July 2025).
Computers 14 00382 g005
Figure 6. Path ranking component.
Figure 6. Path ranking component.
Computers 14 00382 g006
Figure 7. Full example.
Figure 7. Full example.
Computers 14 00382 g007
Figure 8. Hop count distribution for converging, divergent, and linear datasets.
Figure 8. Hop count distribution for converging, divergent, and linear datasets.
Computers 14 00382 g008
Figure 9. Comparison of evaluation results for different hop counts.
Figure 9. Comparison of evaluation results for different hop counts.
Computers 14 00382 g009
Table 1. Comparing existing work with GraphTrace. ✓ = indicates the feature is supported, × = indicates it is not supported.
Table 1. Comparing existing work with GraphTrace. ✓ = indicates the feature is supported, × = indicates it is not supported.
Uses KGEntity ExtractionPath FindingQuery DecompositionSemantic Path RankingContext AggregationLLM-Aware RetrievalModular ArchitectureLatency/Quality TradeoffAnswer Uses KG ContextQuery-to-Path AlignmentAdaptive Multi-Hop ControlSubgraph Diversity HandlingPath CanonicalizationLLM-Based Evaluation StrategyHop-Wise Breakdown
Naive RAG [4]
Advanced RAG [14]() 1() 2
Modular RAG [15]() 3() 2
GraphRAG [9]() 4() 5() 6() 7
LazyGraphRAG [8]() 8() 9() 10() 11() 12
LEGO-GraphRAG [11]() 13() 14() 15
GRAG [16]() 16() 17() 18
GraphTrace (our approach)() 2
Legend: 1: Consists of multiple components but lacks modular or exchangeable units. 2: No explicit mechanism for balancing latency and output quality. 3: Uses scoring mechanisms but lacks semantic path interpretation. 4: Operates on complex queries but does not decompose them explicitly. 5: Retrieval does not consider LLM-specific input constraints or behavior. 6: Uses fixed-hop search; not adaptive to query complexity. 7: No strategy for ensuring diversity among retrieved subgraphs. 8: Performs traversal but not via dedicated path-finding algorithms. 9: Processes entire query without decomposition into subparts. 10: Matches embeddings but lacks explicit segment-to-path alignment. 11: Traversal halts when no improvements occur, but it lacks semantic control. 12: Ranks by score but does not explicitly enforce diversity. 13: Assembles paths from modules but without query-based alignment. 14: Combines modules with variable depth but no dynamic hop adaptation. 15: Combines diverse components but without explicit diversity control. 16: Applies LLM scoring but not deeply semantic (e.g., relation-aware). 17: Allows variable hops but lacks query-driven adaptation. 18: Focuses on best path, not diversity-aware selection.
Table 2. Retrieval evaluation results on the aggregated QA dataset]Retrieval evaluation results on the aggregated question–answer dataset (converging, divergent, and linear). Best results are highlighted in bold. GraphTrace outperformed the baselines in MRR and MAP.
Table 2. Retrieval evaluation results on the aggregated QA dataset]Retrieval evaluation results on the aggregated question–answer dataset (converging, divergent, and linear). Best results are highlighted in bold. GraphTrace outperformed the baselines in MRR and MAP.
Naive
RAG
Naive RAG
+ Subquery
Hybrid
RAG
Rerank
RAG
Naive
GraphRAG
KG
RAG
Graph
TRACE
MRR0.37450.35970.35430.33180.05460.08080.4477
MAP0.08730.11360.08700.08730.01370.01430.1906
Hit@100.79400.82020.79400.84270.15730.17230.8127
Table 3. Retrieval evaluation results on the converging dataset]Retrieval evaluation results on the converging dataset. Best results are highlighted in bold. GraphTrace outperformed the baselines in MRR and MAP.
Table 3. Retrieval evaluation results on the converging dataset]Retrieval evaluation results on the converging dataset. Best results are highlighted in bold. GraphTrace outperformed the baselines in MRR and MAP.
Naive
RAG
Naive RAG
+ Subquery
Hybrid
RAG
Rerank
RAG
Naive
GraphRAG
KG
RAG
Graph
TRACE
MRR0.35020.3680.34310.31020.05810.10130.484
MAP0.08140.13320.08260.0860.01610.01790.2314
Hit@100.78890.87780.76670.83330.26670.25560.8444
Table 4. Retrieval evaluation results on the divergent dataset]Retrieval evaluation results on the divergent dataset. Best results are highlighted in bold. GraphTrace outperformed the baselines in MAP.
Table 4. Retrieval evaluation results on the divergent dataset]Retrieval evaluation results on the divergent dataset. Best results are highlighted in bold. GraphTrace outperformed the baselines in MAP.
Naive
RAG
Naive RAG
+ Subquery
Hybrid
RAG
Rerank
RAG
Naive
GraphRAG
KG
RAG
Graph
TRACE
MRR0.42610.37660.39250.35510.0170.08630.4149
MAP0.10380.11640.09950.09260.00490.01440.175
Hit@100.80.83330.81110.84440.05560.16670.8
Table 5. Retrieval evaluation results on the linear dataset]Retrieval evaluation results on the linear dataset. Best results are highlighted in bold. GraphTrace outperformed the baselines in MRR and MAP.
Table 5. Retrieval evaluation results on the linear dataset]Retrieval evaluation results on the linear dataset. Best results are highlighted in bold. GraphTrace outperformed the baselines in MRR and MAP.
Naive
RAG
Naive RAG
+ Subquery
Hybrid
RAG
Rerank
RAG
Naive
GraphRAG
KG
RAG
Graph
TRACE
MRR0.34640.33360.32620.330.08970.0540.4442
MAP0.07630.09050.07850.08310.02030.01030.1644
Hit@100.79310.74710.80460.85060.14940.0920.7931
Table 6. Retrieval performance in the converging dataset, broken down by hop count. The table reports the MRR, MAP, and Hit@10 across all methods for questions requiring between four and eight reasoning hops. Best results are highlighted in bold.
Table 6. Retrieval performance in the converging dataset, broken down by hop count. The table reports the MRR, MAP, and Hit@10 across all methods for questions requiring between four and eight reasoning hops. Best results are highlighted in bold.
Hop CountQ/A CountNaive RAGNaive RAG + SubqueryHybrid RAGRerank RAGNaive GraphRAGKG RAGGraphTrace
MRR
410.50000.50000.33330.50000.00000.00000.2000
520.75000.26670.37500.50000.00000.00000.7500
6230.40940.46500.53450.43010.06050.03360.5765
7470.34820.35010.30550.28890.06070.16060.5377
8170.21960.29080.18490.17370.05800.04700.1960
MAP
410.12500.25000.08330.20830.00000.00000.1056
520.19000.15170.10830.24290.00000.00000.3100
6230.12500.20730.15970.13660.01870.00720.3276
7470.06770.10780.05910.07140.01450.02850.2433
8170.04520.09400.04020.03220.01960.00640.0667
Hit@10
411.00001.00001.00001.00000.00000.00001.0000
521.00001.00001.00001.00000.00000.00001.0000
6230.95650.91300.95650.95650.39130.17391.0000
7470.76600.89360.74470.85110.21280.36170.8511
8170.58820.76470.52940.58820.29410.11760.5882
Table 7. Retrieval performance in the divergent dataset, broken down by hop count. The table reports the MRR, MAP, and Hit@10 across all methods for questions requiring between four and eight reasoning hops. Best results are highlighted in bold.
Table 7. Retrieval performance in the divergent dataset, broken down by hop count. The table reports the MRR, MAP, and Hit@10 across all methods for questions requiring between four and eight reasoning hops. Best results are highlighted in bold.
Hop CountQ/A CountNaive RAGNaive RAG + SubqueryHybrid RAGRerank RAGNaive GraphRAGKG RAGGraphTrace
MRR
410.25000.50000.25000.25000.00000.00000.2000
5100.29170.20100.36170.30170.03240.10910.3169
6700.44330.40370.39930.36890.01480.07980.4356
780.39370.26610.37620.29580.01040.01140.3721
811.00001.00000.50000.50000.09091.00000.5000
MAP
410.06250.12500.06250.06250.00000.00000.2339
5100.10100.08130.13100.09130.00900.02180.1377
6700.10600.12330.09850.09430.00350.01330.1797
780.08360.07040.07160.07030.00150.00310.1650
810.17500.34380.11250.19940.09680.12500.2396
Hit@10
411.00001.00001.00001.00000.00000.00001.0000
5100.70000.80000.70000.90000.10000.10000.8000
6700.80000.81430.81430.81430.05710.18570.7857
780.87501.00000.87501.00000.00000.00000.8750
811.00001.00001.00001.00000.00001.00001.0000
Table 8. Retrieval performance in linear dataset by hop count] Retrieval performance in the linear dataset, broken down by hop count. The table reports the MRR, MAP, and Hit@10 across all methods for questions requiring between four and six reasoning hops. Best results are highlighted in bold.
Table 8. Retrieval performance in linear dataset by hop count] Retrieval performance in the linear dataset, broken down by hop count. The table reports the MRR, MAP, and Hit@10 across all methods for questions requiring between four and six reasoning hops. Best results are highlighted in bold.
Hop CountQ/A CountNaive RAGNaive RAG + SubqueryHybrid RAGRerank RAGNaive GraphRAGKG RAGGraphTrace
MRR
440.25000.35420.15420.32080.06250.00000.4750
560.36670.27500.31110.34720.08330.16670.3806
6770.34980.33710.33640.32920.09160.04800.4475
MAP
440.06250.15220.03850.15760.01560.00000.3552
560.09780.06450.07170.08850.01670.03330.1083
6770.07540.08930.08110.07880.02080.00900.1589
Hit@10
441.00001.00000.75001.00000.25000.00001.0000
560.66670.66670.66670.83330.16670.16670.8333
6770.79220.74030.81820.84420.14290.09090.7792
Table 9. Generation performance in the converging, divergent, and linear question–answer datasets. Best results are highlighted in bold.
Table 9. Generation performance in the converging, divergent, and linear question–answer datasets. Best results are highlighted in bold.
CriterionNaive RAGNaive RAG + SubqueryHybrid RAGRerank RAGNaive GraphRAGKG RAGGraphTrace
Converging
Comprehensiveness3.70%16.30%0.37%15.93%30.74%12.96%20.00%
Diversity6.30%26.67%3.33%16.30%15.93%20.37%11.11%
Empowerment2.59%17.04%2.22%13.33%26.30%20.74%17.78%
Directness6.30%19.26%4.07%12.22%3.70%23.70%30.74%
Overall3.70%17.04%0.37%15.93%30.37%12.59%20.00%
Divergent
Comprehensiveness13.33%10.74%5.56%14.44%20.37%15.93%19.63%
Diversity11.11%24.81%5.56%13.33%10.00%16.67%18.52%
Empowerment13.33%10.74%4.07%13.33%18.89%21.85%17.78%
Directness19.26%11.11%8.89%14.07%0.37%17.04%29.26%
Overall13.70%10.74%5.56%14.81%19.63%15.56%20.00%
Linear
Comprehensiveness19.54%11.88%2.68%17.62%17.62%14.94%15.71%
Diversity15.71%22.99%7.28%12.26%15.33%9.58%16.86%
Empowerment18.39%11.88%2.68%14.94%18.01%18.77%15.33%
Directness21.07%8.81%2.30%15.71%5.75%22.22%24.14%
Overall18.39%13.41%2.30%17.24%17.24%15.71%15.71%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Osipjan, A.; Khorashadizadeh, H.; Kessel, A.-L.; Groppe, S.; Groppe, J. GraphTrace: A Modular Retrieval Framework Combining Knowledge Graphs and Large Language Models for Multi-Hop Question Answering. Computers 2025, 14, 382. https://doi.org/10.3390/computers14090382

AMA Style

Osipjan A, Khorashadizadeh H, Kessel A-L, Groppe S, Groppe J. GraphTrace: A Modular Retrieval Framework Combining Knowledge Graphs and Large Language Models for Multi-Hop Question Answering. Computers. 2025; 14(9):382. https://doi.org/10.3390/computers14090382

Chicago/Turabian Style

Osipjan, Anna, Hanieh Khorashadizadeh, Akasha-Leonie Kessel, Sven Groppe, and Jinghua Groppe. 2025. "GraphTrace: A Modular Retrieval Framework Combining Knowledge Graphs and Large Language Models for Multi-Hop Question Answering" Computers 14, no. 9: 382. https://doi.org/10.3390/computers14090382

APA Style

Osipjan, A., Khorashadizadeh, H., Kessel, A.-L., Groppe, S., & Groppe, J. (2025). GraphTrace: A Modular Retrieval Framework Combining Knowledge Graphs and Large Language Models for Multi-Hop Question Answering. Computers, 14(9), 382. https://doi.org/10.3390/computers14090382

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop