Rewind-IL: Online Failure Detection and State Respawning for Imitation Learning
arXiv cs.RO / 4/21/2026
📰 NewsDeveloper Stack & InfrastructureModels & Research
Key Points
- Rewind-IL is presented as a training-free online safeguard for imitation learning systems that use generative, action-chunked policies, targeting reliability issues when execution drifts away from demonstration behavior.
- It uses a zero-shot failure detector based on TIDE (Temporal Inter-chunk Discrepancy Estimate) and calibrates decisions with split conformal prediction to reduce false triggers under benign changes.
- When a failure is detected, Rewind-IL employs a “state respawning” mechanism that rewinds the robot to a semantically verified safe intermediate state and then restarts inference from a clean policy state.
- The approach builds an offline recovery-checkpoint library using a vision-language model over demonstrations, then matches online execution to checkpoint features via a compact database constructed from a frozen policy encoder.
- Experiments on long-horizon real and simulated manipulation tasks (including transfer to flow-matching action-chunked policies) indicate improved robustness by combining internal policy consistency checks with semantics-grounded recovery.
Related Articles

A practical guide to getting comfortable with AI coding tools
Dev.to

We built it during the NVIDIA DGX Spark Full-Stack AI Hackathon — and it ended up winning 1st place overall 🏆
Dev.to

Stop Losing Progress: Setting Up a Pro Jupyter Workflow in VS Code (No More Colab Timeouts!)
Dev.to

🚀 Major BrowserAct CLI Update
Dev.to

Building AgentOS: Why I’m Building the AWS Lambda for Insurance Claims
Dev.to