S-Path-RAG: Semantic-Aware Shortest-Path Retrieval Augmented Generation for Multi-Hop Knowledge Graph Question Answering

arXiv cs.CL / 2026/3/26

📰 ニュースSignals & Early TrendsIdeas & Deep AnalysisModels & Research

要点

  • S-Path-RAG is a semantic-aware Retrieval-Augmented Generation framework that improves multi-hop question answering on large knowledge graphs by enumerating bounded-length, semantically weighted candidate paths rather than relying on one-shot, text-heavy retrieval.
  • The approach combines a hybrid weighted k-shortest/beam search with constrained random walks, learning a differentiable path scorer and using a contrastive path encoder plus a lightweight verifier to improve path selection.
  • It injects a compact soft mixture of selected path latents into a language model through cross-attention, aiming to be token-efficient while remaining topology-aware.
  • S-Path-RAG operates in an iterative “Neural-Socratic Graph Dialogue” loop where the language model produces concise diagnostics that are mapped to targeted graph edits or seed expansions for adaptive retrieval under uncertainty.
  • Experiments on standard multi-hop KGQA benchmarks show consistent gains in answer accuracy, evidence coverage, and end-to-end efficiency, with analyses and deployment recommendations for constrained compute and token budgets.

Abstract

We present S-Path-RAG, a semantic-aware shortest-path Retrieval-Augmented Generation framework designed to improve multi-hop question answering over large knowledge graphs. S-Path-RAG departs from one-shot, text-heavy retrieval by enumerating bounded-length, semantically weighted candidate paths using a hybrid weighted k-shortest, beam, and constrained random-walk strategy, learning a differentiable path scorer together with a contrastive path encoder and lightweight verifier, and injecting a compact soft mixture of selected path latents into a language model via cross-attention. The system runs inside an iterative Neural-Socratic Graph Dialogue loop in which concise diagnostic messages produced by the language model are mapped to targeted graph edits or seed expansions, enabling adaptive retrieval when the model expresses uncertainty. This combination yields a retrieval mechanism that is both token-efficient and topology-aware while preserving interpretable path-level traces for diagnostics and intervention. We validate S-Path-RAG on standard multi-hop KGQA benchmarks and through ablations and diagnostic analyses. The results demonstrate consistent improvements in answer accuracy, evidence coverage, and end-to-end efficiency compared to strong graph- and LLM-based baselines. We further analyze trade-offs between semantic weighting, verifier filtering, and iterative updates, and report practical recommendations for deployment under constrained compute and token budgets.