Breaking Lock-In: Preserving Steerability under Low-Data VLA Post-Training
arXiv cs.RO / 4/28/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- The paper studies “lock-in” in vision-language-action (VLA) policies after low-data supervised fine-tuning, where the model becomes overly specialized and stops handling novel instructions.
- It characterizes two failure modes—concept lock-in (over-fixation on training objects/attributes) and spatial lock-in (over-fixation on training spatial targets).
- The authors propose DeLock, which mitigates lock-in by preserving visual grounding during post-training and using test-time contrastive prompt guidance to steer the policy’s denoising dynamics.
- Across eight simulation and real-world evaluations, DeLock outperforms strong baselines and can match or exceed the performance of a state-of-the-art generalist VLA post-trained with much more curated data.
- The approach reduces the need for extra supervision signals or augmented datasets by leveraging the model’s internal pre-trained knowledge during post-training.
💡 Insights using this article
This article is featured in our daily AI news digest — key takeaways and action items at a glance.
Related Articles

Big Tech firms are accelerating AI investments and integration, while regulators and companies focus on safety and responsible adoption.
Dev.to

Same Agent, Different Risk | How Microsoft 365 Copilot Grounding Changes the Security Model | Rahsi Framework™
Dev.to

Claude Haiku for Low-Cost AI Inference: Patterns from a Horse Racing Prediction System
Dev.to

How We Built an Ambient AI Clinical Documentation Pipeline (and Saved Doctors 8+ Hours a Week)
Dev.to

🦀 PicoClaw Deep Dive — A Field Guide to Building an Ultra-Light AI Agent in Go 🐹
Dev.to