MotuBrain: An Advanced World Action Model for Robot Control

arXiv cs.RO / 5/1/2026

📰 NewsDeveloper Stack & InfrastructureModels & Research

Key Points

  • MotuBrain is a new Vision-Language-Action (VLA) world action model designed to better capture fine-grained world dynamics for robot control.
  • The model unifies video and action modeling using a UniDiffuser formulation with a three-stream Mixture-of-Transformers architecture.
  • A single MotuBrain model can run in multiple inference modes, including policy learning, world modeling, video generation, inverse dynamics, and joint video-action prediction.
  • It is built to scale across heterogeneous multimodal datasets, including video-only data and cross-embodiment robot data.
  • For real-world deployment, MotuBrain adds unified multiview representations and explicit language-action coupling, along with an efficient inference stack that reportedly delivers 50x+ speedups for real-time use.

Abstract

Vision-Language-Action (VLA) models achieve strong semantic generalization but often lack fine-grained modeling of world dynamics. Recent work explores video generation models as a foundation for world modeling, leading to unified World Action Models (WAMs) that jointly model visual dynamics and actions. We present MotuBrain, a unified multimodal generative model that jointly models video and action under a UniDiffuser formulation with a three-stream Mixture-of-Transformers architecture. A single model supports multiple inference modes, including policy learning, world modeling, video generation, inverse dynamics, and joint video-action prediction, while scaling to heterogeneous multimodal data such as video-only and cross-embodiment robot data. To improve real-world applicability, MotuBrain introduces a unified multiview representation, explicit language-action coupling, and an efficient inference stack, achieving over 50x speedup for real-time deployment.