🤖 AI Summary
Existing GraphRAG approaches struggle to balance accuracy and efficiency in multi-hop reasoning due to the high computational cost of graph traversal and semantic noise introduced by LLM-based summarization. To address these limitations, this work proposes the HyperNode abstraction mechanism, which constructs structured reasoning paths through iterative knowledge-triplet-based path building and employs a logic-path-guided evidence localization strategy to directly map to relevant source text segments. This approach eliminates the need for random walks and avoids semantic distortion while preserving knowledge completeness. Experimental results demonstrate that the method significantly enhances retrieval efficiency and achieves superior performance across multiple single-hop and multi-hop question answering benchmarks, yielding up to a 28.8× speedup over state-of-the-art GraphRAG methods.
📝 Abstract
Large Language Models (LLMs) often struggle with inherent knowledge boundaries and hallucinations, limiting their reliability in knowledge-intensive tasks. While Retrieval-Augmented Generation (RAG) mitigates these issues, it frequently overlooks structural interdependencies essential for multi-hop reasoning. Graph-based RAG approaches attempt to bridge this gap, yet they typically face trade-offs between accuracy and efficiency due to challenges such as costly graph traversals and semantic noise in LLM-generated summaries. In this paper, we propose HyperNode Expansion and Logical Path-Guided Evidence Localization strategies for GraphRAG (HELP), a novel framework designed to balance accuracy with practical efficiency through two core strategies: 1) HyperNode Expansion, which iteratively chains knowledge triplets into coherent reasoning paths abstracted as HyperNodes to capture complex structural dependencies and ensure retrieval accuracy; and 2) Logical Path-Guided Evidence Localization, which leverages precomputed graph-text correlations to map these paths directly to the corpus for superior efficiency. HELP avoids expensive random walks and semantic distortion, preserving knowledge integrity while drastically reducing retrieval latency. Extensive experiments demonstrate that HELP achieves competitive performance across multiple simple and multi-hop QA benchmarks and up to a 28.8$\times$ speedup over leading Graph-based RAG baselines.