GenerativeMPC: VLM-RAG-guided Whole-Body MPC with Virtual Impedance for Bimanual Mobile Manipulation
arXiv cs.RO / 4/22/2026
📰 NewsDeveloper Stack & InfrastructureModels & Research
Key Points
- The paper introduces GenerativeMPC, a hierarchical cyber-physical framework that connects semantic scene understanding to physical control constraints for bimanual mobile manipulation.
- It uses a Vision-Language Model with Retrieval-Augmented Generation (VLM-RAG) to convert visual and linguistic context into grounded MPC outputs such as dynamic velocity limits and safety margins.
- The same VLM-RAG component adjusts virtual stiffness and damping gains for a unified impedance-admittance controller, allowing context-aware compliant behavior during human-robot interaction.
- An experience-driven vector database is used to maintain consistent parameter grounding without retraining, improving practicality and stability.
- Experiments in MuJoCo, IsaacSim, and on a physical bimanual platform show about a 60% speed reduction near humans and demonstrate safe, socially aware navigation and manipulation.
Related Articles
Autoencoders and Representation Learning in Vision
Dev.to
Every AI finance app wants your data. I didn’t trust that — so I built my own. Offline.
Dev.to
Control Claude with Just a URL. The Chrome Extension "Send to Claude" Is Incredibly Useful
Dev.to
Google Stitch 2.0: Senior-Level UI in Seconds, But Editing Still Breaks
Dev.to

Now Meta will track what employees do on their computers to train its AI agents
The Verge