UltRAG: a Universal Simple Scalable Recipe for Knowledge Graph RAG

arXiv cs.CL / 4/1/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces ULTRAG, a universal framework for applying retrieval-augmented generation to knowledge graphs, targeting multi-hop and multi-node question answering where classical KG-RAG is difficult.
  • ULTRAG improves KGQA by equipping off-the-shelf LLMs with neural query-executing modules, enabling graph querying with LLMs rather than relying on retraining.
  • The authors report that ULTRAG attains state-of-the-art performance on KGQA benchmarks compared with prior KG-RAG approaches.
  • Results suggest ULTRAG can connect to Wikidata-scale graphs (116M entities, 1.6B relations) at comparable or lower computational costs than existing solutions.

Abstract

Large language models (LLMs) frequently generate confident yet factually incorrect content when used for language generation (a phenomenon often known as hallucination). Retrieval augmented generation (RAG) tries to reduce factual errors by identifying information in a knowledge corpus and putting it in the context window of the model. While this approach is well-established for document-structured data, it is non-trivial to adapt it for Knowledge Graphs (KGs), especially for queries that require multi-node/multi-hop reasoning on graphs. We introduce ULTRAG, a general framework for retrieving information from Knowledge Graphs that shifts away from classical RAG. By endowing LLMs with off-the-shelf neural query executing modules, we highlight how readily available language models can achieve state-of-the-art results on Knowledge Graph Question Answering (KGQA) tasks without any retraining of the LLM or executor involved. In our experiments, ULTRAG achieves better performance when compared to state-of-the-art KG-RAG solutions, and it enables language models to interface with Wikidata-scale graphs (116M entities, 1.6B relations) at comparable or lower costs.