UnWeaving the knots of GraphRAG -- turns out VectorRAG is almost enough

arXiv cs.CL / 4/1/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that chunk-level VectorRAG treats retrieved chunks as atomic vectors, which makes it difficult to support multi-hop questions and preserve structured relations.
  • It critiques GraphRAG approaches for requiring graph-structured indices with much higher complexity and often relying on retrieval heuristics.
  • The proposed framework, UnWeaver, uses an LLM to decompose documents into cross-chunk entities, then uses these entities to retrieve the original text chunks during RAG to maintain fidelity.
  • The authors claim entity-based decomposition can reduce indexing/generation noise while simplifying GraphRAG’s benefits without building full graph indices.
  • Overall, the work positions “VectorRAG being almost enough” by showing a middle ground between pure vector retrieval and full knowledge-graph RAG.

Abstract

One of the key problems in Retrieval-augmented generation (RAG) systems is that chunk-based retrieval pipelines represent the source chunks as atomic objects, mixing the information contained within such a chunk into a single vector. These vector representations are then fundamentally treated as isolated, independent and self-sufficient, with no attempt to represent possible relations between them. Such an approach has no dedicated mechanisms for handling multi-hop questions. Graph-based RAG systems aimed to ameliorate this problem by modeling information as knowledge-graphs, with entities represented by nodes being connected by robust relations, and forming hierarchical communities. This approach however suffers from its own issues with some of them being: orders of magnitude increased componential complexity in order to create graph-based indices, and reliance on heuristics for performing retrieval. We propose UnWeaver, a novel RAG framework simplifying the idea of GraphRAG. UnWeaver disentangles the contents of the documents into entities which can occur across multiple chunks using an LLM. In the retrieval process entities are used as an intermediate way of recovering original text chunks hence preserving fidelity to the source material. We argue that entity-based decomposition yields a more distilled representation of original information, and additionally serves to reduce noise in the indexing, and generation process.