TSUBASA: Improving Long-Horizon Personalization via Evolving Memory and Self-Learning with Context Distillation
arXiv cs.CL / 4/10/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- TSUBASA is proposed as a two-part method to improve personalized LLM performance on long-horizon tasks by evolving how user information is written to memory and how it is read back.
- The approach addresses key weaknesses in prior memory mechanisms, including difficulty tracking evolving user behavior over long conversation/activity histories.
- TSUBASA also targets the RAG quality–efficiency tradeoff and the train–inference gap in parametric adaptation by using a self-learning objective with context distillation to internalize user experiences.
- Experiments on long-horizon benchmarks with the Qwen-3 model family (4B–32B) show TSUBASA outperforms memory-augmented competitors like Mem0 and Memory-R1, which rely more heavily on memory writing.
- The authors report Pareto improvements that deliver robust, high-fidelity personalization while reducing token budget compared with prior approaches.
💡 Insights using this article
This article is featured in our daily AI news digest — key takeaways and action items at a glance.
Related Articles

Black Hat Asia
AI Business

GLM 5.1 tops the code arena rankings for open models
Reddit r/LocalLLaMA

Big Tech firms are accelerating AI investments and integration, while regulators and companies focus on safety and responsible adoption.
Dev.to

My Bestie Built a Free MCP Server for Job Search — Here's How It Works
Dev.to
can we talk about how AI has gotten really good at lying to you?
Reddit r/artificial