InfiniteDance: Scalable 3D Dance Generation Towards in-the-wild Generalization
arXiv cs.CV / 3/17/2026
📰 NewsSignals & Early TrendsModels & Research
Key Points
- The work tackles generalizable 3D dance generation from monocular videos by scaling both data and model design to handle unseen music.
- It introduces the Foot Restoration Diffusion Model (FRDM) to enforce physical plausibility via foot-contact and geometric constraints, creating a 100.69-hour multimodal 3D dance dataset.
- It presents ChoreoLLaMA, a scalable LLaMA-based architecture with a retrieval-augmented generation module to reference dance prompts under unfamiliar music conditions.
- A slow/fast-cadence Mixture-of-Experts module lets the model adapt motion rhythms across different tempos.
- Experiments show improvements over existing methods, and the authors plan to release code, models, and data.




