RISE: Self-Improving Robot Policy with Compositional World Model
arXiv cs.RO / 4/29/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper introduces RISE, a framework for robotic reinforcement learning that aims to make Vision-Language-Action (VLA) policies more robust in contact-rich, dynamic manipulation tasks where small execution errors can cascade into failures.
- RISE uses a Compositional World Model with a controllable dynamics component to predict multi-view future states and a progress/value model to score imagined outcomes and compute informative advantages.
- By separating the world-model and value/evaluation components, the approach tailors state prediction and value estimation using distinct architectures aligned to different objectives.
- The system runs a closed-loop “self-improving” pipeline that repeatedly generates imagined rollouts, estimates advantages, and updates the policy entirely in imaginary space, reducing the need for expensive and risky on-policy physical RL.
- Experiments on three real-world tasks show substantial gains over prior work, including more than +35% absolute improvement in dynamic brick sorting, +45% in backpack packing, and +35% in box closing.
Related Articles
LLMs will be a commodity
Reddit r/artificial

Indian Developers: How to Build AI Side Income with $0 Capital in 2026
Dev.to

What it feels like to have to have Qwen 3.6 or Gemma 4 running locally
Reddit r/LocalLLaMA

Dex lands $5.3M to grow its AI-driven talent matching platform
Tech.eu

AI Citation Registry: Why Daily Updates Leave No Time for Data Structuring
Dev.to