AI Navigate

Multi-hop Reasoning and Retrieval in Embedding Space: Leveraging Large Language Models with Knowledge

arXiv cs.AI / 3/17/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces EMBRAG, an embedding-based retrieval and reasoning framework that enhances LLMs by leveraging knowledge graphs to improve reasoning reliability and reduce hallucinations.
  • It first generates multiple logical rules grounded in the knowledge graph from the input query, then applies these rules in the embedding space to guide reasoning with KG guidance.
  • A reranker model is used to interpret and refine the generated rules, improving robustness against incomplete or noisy KG data.
  • Experimental results on two KGQA benchmarks demonstrate state-of-the-art performance in KG reasoning tasks.

Abstract

As large language models (LLMs) continue to grow in size, their abilities to tackle complex tasks have significantly improved. However, issues such as hallucination and the lack of up-to-date knowledge largely remain unresolved. Knowledge graphs (KGs), which serve as symbolic representations of real-world knowledge, offer a reliable source for enhancing reasoning. Integrating KG retrieval into LLMs can therefore strengthen their reasoning by providing dependable knowledge. Nevertheless, due to limited understanding of the underlying knowledge graph, LLMs may struggle with queries that have multiple interpretations. Additionally, the incompleteness and noise within knowledge graphs may result in retrieval failures. To address these challenges, we propose an embedding-based retrieval reasoning framework EMBRAG. In this approach, the model first generates multiple logical rules grounded in knowledge graphs based on the input query. These rules are then applied to reasoning in the embedding space, guided by the knowledge graph, ensuring more robust and accurate reasoning. A reranker model further interprets these rules and refines the results. Extensive experiments on two benchmark KGQA datasets demonstrate that our approach achieves the new state-of-the-art performance in KG reasoning tasks.