MoRight: Motion Control Done Right
arXiv cs.CV / 4/9/2026
📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- MoRight is presented as a new unified framework for generating motion-controlled videos where user actions produce physically plausible scene dynamics from freely chosen viewpoints.
- The method improves on prior work by disentangling object motion from camera motion using canonical-view motion specification and temporal cross-view attention for transfer to target viewpoints.
- MoRight explicitly models motion causality by decomposing motion into active (user-driven) and passive (consequence) components, learning how non-actuated objects react coherently rather than only translating pixels.
- It supports both forward reasoning (predict consequences from supplied active motion) and inverse reasoning (recover plausible driving actions from desired passive outcomes), while retaining viewpoint freedom.
- Experiments on three benchmarks report state-of-the-art results across generation quality, motion controllability, and interaction awareness.
💡 Insights using this article
This article is featured in our daily AI news digest — key takeaways and action items at a glance.
Related Articles

Black Hat Asia
AI Business

Amazon CEO takes aim at Nvidia, Intel, Starlink, more in annual shareholder letter
TechCrunch

Why Anthropic’s new model has cybersecurity experts rattled
Reddit r/artificial
Does the AI 2027 paper still hold any legitimacy?
Reddit r/artificial

Why Most Productivity Systems Fail (And What to Do Instead)
Dev.to