SGA-MCTS: Decoupling Planning from Execution via Training-Free Atomic Experience Retrieval
arXiv cs.AI / 4/17/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- SGA-MCTS reframes LLM multi-step planning as a non-parametric retrieval problem to avoid the trade-off between slow inference-time search and limited generalization from supervised fine-tuning.
- The method uses offline Monte Carlo Tree Search to generate high-fidelity solution trajectories, then distills them into reusable State-Goal-Action (SGA) atoms that are de-lexicalized to abstract away domain-specific details.
- At inference time, a retrieval-augmented hybrid symbolic-semantic agent fetches relevant SGAs and re-grounds them into the current context as soft reasoning hints, improving planning without heavy online search.
- Experiments on complex benchmarks report that frozen, open-weights models using SGA-MCTS can reach performance comparable to state-of-the-art systems (e.g., GPT-5) without task-specific fine-tuning.
- By amortizing expensive search costs offline, SGA-MCTS aims to deliver “System 2” reasoning depth at “System 1” inference speed, making real-time autonomous planning more scalable.


![[2026] OpenTelemetry for LLM Observability — Self-Hosted Setup](/_next/image?url=https%3A%2F%2Fmedia2.dev.to%2Fdynamic%2Fimage%2Fwidth%3D1200%2Cheight%3D627%2Cfit%3Dcover%2Cgravity%3Dauto%2Cformat%3Dauto%2Fhttps%253A%252F%252Fdev-to-uploads.s3.amazonaws.com%252Fuploads%252Farticles%252Flu4b6ttuhur71z5gemm0.png&w=3840&q=75)
