KG-Hopper: Empowering Compact Open LLMs with Knowledge Graph Reasoning via Reinforcement Learning
arXiv cs.CL / 3/24/2026
📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper introduces KG-Hopper, a reinforcement learning framework designed to improve compact open LLMs on knowledge graph multi-hop question answering by reducing brittle step-by-step pipelines.
- Instead of executing reasoning in isolated sequential steps, KG-Hopper trains a 7B “Reasoning LLM” to embed the full knowledge-graph traversal and decision process into a single unified thinking stage with dynamic path exploration and backtracking.
- Experiments across eight KG reasoning benchmarks show KG-Hopper outperforms larger multi-step systems (up to 70B) and matches competitive performance against proprietary models like GPT-3.5-Turbo and GPT-4o-mini.
- The approach is reported to be compact, open, and data-efficient, and the authors provide public code via the linked GitHub repository.
Related Articles

Black Hat Asia
AI Business

"The Agent Didn't Decide Wrong. The Instructions Were Conflicting — and Nobody Noticed."
Dev.to
Top 5 LLM Gateway Alternatives After the LiteLLM Supply Chain Attack
Dev.to

Stop Counting Prompts — Start Reflecting on AI Fluency
Dev.to

Reliable Function Calling in Deeply Recursive Union Types: Fixing Qwen Models' Double-Stringify Bug
Dev.to