PhyCo: Learning Controllable Physical Priors for Generative Motion
arXiv cs.CV / 5/1/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- PhyCo is a new framework aimed at improving physical consistency in video diffusion models, addressing issues like object drift, unrealistic collisions, and mismatched material responses.
- It combines a large photorealistic simulation dataset (100K+ videos with systematically varied friction, restitution, deformation, and forces), physics-supervised fine-tuning using a ControlNet conditioned on pixel-aligned physical property maps, and VLM-guided reward optimization via physics-targeted queries.
- The method enables controllable, physically plausible motion generation without requiring a simulator or geometry reconstruction at inference time.
- Experiments on the Physics-IQ benchmark show substantial gains in physical realism over strong baselines, and human evaluations indicate more faithful control over physical attributes.
- Overall, the work proposes a scalable route to physically consistent and controllable generative video models that can generalize beyond the synthetic environments used for training.
Related Articles

Why Autonomous Coding Agents Keep Failing — And What Actually Works
Dev.to

Why Enterprise AI Pilots Fail
Dev.to

The PDF Feature Nobody Asked For (That I Use Every Day)
Dev.to

How to Fix OpenClaw Tool Calling Issues
Dev.to

Mistral's new flagship Medium 3.5 folds chat, reasoning, and code into one model
THE DECODER