MetaClaw: Just Talk -- An Agent That Meta-Learns and Evolves in the Wild
arXiv cs.LG / 3/19/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- MetaClaw introduces a continual meta-learning framework that jointly evolves a base LLM policy and a library of reusable skills to adapt to shifting user needs without downtime.
- It combines skill-driven fast adaptation, which synthesizes new skills from failure trajectories via an LLM evolver, with opportunistic policy optimization using cloud LoRA fine-tuning and RL with a Process Reward Model, triggered during user-inactive windows by the Opportunistic Meta-Learning Scheduler.
- The approach uses a versioning mechanism to separate support and query data and a proxy-based architecture that scales production-size LLMs without local GPUs, enabling deployment in real workloads.
- Empirical results on MetaClaw-Bench and AutoResearchClaw show up to 32% relative accuracy gains and improvements from 21.4% to 40.6% on Kimi-K2.5, with an 18.3% increase in composite robustness; the code is available at GitHub.
Related Articles
[R] Combining Identity Anchors + Permission Hierarchies achieves 100% refusal in abliterated LLMs — system prompt only, no fine-tuning
Reddit r/MachineLearning
How I Built an AI SDR Agent That Finds Leads and Writes Personalized Cold Emails
Dev.to
Complete Guide: How To Make Money With Ai
Dev.to
I Analyzed My Portfolio with AI and Scored 53/100 — Here's How I Fixed It to 85+
Dev.to
The Demethylation
Dev.to