DanceCrafter: Fine-Grained Text-Driven Controllable Dance Generation via Choreographic Syntax
arXiv cs.AI / 4/22/2026
📰 NewsDeveloper Stack & InfrastructureModels & Research
Key Points
- The paper proposes a new theoretical framework called “Choreographic Syntax” to better describe and annotate complex, text-driven controllable dance instructions.
- It builds “DanceFlow,” a highly fine-grained dataset combining professional dance archives with high-fidelity motion capture, totaling 41 hours of motion and 6.34 million words of descriptions.
- It introduces “DanceCrafter,” a tailored motion-transformer model based on the Momentum Human Rig, using a continuous manifold motion representation and hybrid normalization to improve training stability.
- The model also uses an anatomy-aware loss to regulate the natural decoupled movement of different body parts, enabling stable and high-fidelity dance generation.
- Extensive evaluations and user studies report state-of-the-art results in motion quality, fine-grained controllability, and naturalness of generated sequences.
Related Articles

GeoReg LLM-Driven Few-Shot Socio-Economic Estimation for Data-Scarce Regions
Dev.to

Enterprise AI Governance Has Shifted from Policy to Execution
Dev.to

Big Tech firms are accelerating AI investments and integration, while regulators and companies focus on safety and responsible adoption.
Dev.to

Rethinking CNN Models for Audio Classification
Dev.to

Database world trying to build natural language query systems again – this time with LLMs
The Register