AnimateAnyMesh++: A Flexible 4D Foundation Model for High-Fidelity Text-Driven Mesh Animation

arXiv cs.CV / 4/30/2026

📰 NewsSignals & Early TrendsModels & Research

Key Points

  • The paper introduces AnimateAnyMesh++, a feed-forward foundation model designed to generate high-fidelity 3D mesh animations from text for arbitrary input meshes despite limited 4D training data.
  • It expands the DyMesh-XL dataset by mining dynamic content from Objaverse-XL, increasing unique identities from 60K to 300K and boosting diversity in categories and motions.
  • The authors upgrade DyMeshVAE-Flex with power-law topology-aware attention and vertex-normal enhanced features to improve trajectory reconstruction, preserve local geometry, and reduce “trajectory sticking” artifacts.
  • They modify both DyMeshVAE-Flex and the rectified-flow (RF) generator to support variable-length sequence training and generation, enabling longer animations while maintaining reconstruction quality.
  • Experiments show the method produces semantically accurate and temporally coherent mesh animations in seconds and outperforms prior approaches, with expected gains across benchmarks and real-world meshes, along with plans to release code and models.

Abstract

Recent advances in 4D content generation have attracted increasing attention, yet creating high-quality animated 3D models remains challenging due to the complexity of modeling spatio-temporal distributions and the scarcity of 4D training data. We present AnimateAnyMesh++, a feed-forward framework for text-driven animation of arbitrary 3D meshes with substantial upgrades in data, architecture, and generative capability. First, we expand the DyMesh-XL dataset by mining dynamic content from Objaverse-XL, increasing the number of unique identities from 60K to 300K and substantially broadening category and motion diversity. Second, we redesign DyMeshVAE-Flex with power-law topology-aware attention and vertex-normal enhanced features, which significantly improves trajectory reconstruction, local geometry preservation, and mitigates trajectory-sticking artifacts. Third, we introduce architectural changes to both DyMeshVAE-Flex and the rectified-flow (RF) generator to support variable-length sequence training and generation, enabling longer animations while preserving reconstruction fidelity. Extensive experiments demonstrate that AnimateAnyMesh++ generates semantically accurate and temporally coherent mesh animations within seconds, surpassing prior approaches in quality and efficiency. The enlarged DyMesh-XL, the upgraded DyMeshVAE-Flex, and variable-length RF together deliver consistent gains across benchmarks and in-the-wild meshes. We will release code, models, and the expanded DyMesh-XL upon acceptance of this manuscript to facilitate research in 4D content creation.