Polyglot: Multilingual Style Preserving Speech-Driven Facial Animation

arXiv cs.CV / 4/20/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper tackles multilingual Speech-Driven Facial Animation (SDFA), noting that prior models trained on single-language data struggle in real-world multilingual use.
  • It introduces Polyglot, a unified diffusion-based architecture that jointly models both language (via transcript embeddings) and individual speaking style (via style embeddings from reference facial sequences).
  • The approach avoids the need for predefined language or speaker labels by relying on self-supervised learning, aiming to generalize across languages and speakers.
  • Experiments indicate improved performance in both monolingual and multilingual settings, with generated animations better capturing rhythm, articulation, intonation-related expression, and habitual facial movements.
  • By conditioning simultaneously on language and personal style, Polyglot produces more temporally coherent and realistic facial animations driven by speech.

Abstract

Speech-Driven Facial Animation (SDFA) has gained significant attention due to its applications in movies, video games, and virtual reality. However, most existing models are trained on single-language data, limiting their effectiveness in real-world multilingual scenarios. In this work, we address multilingual SDFA, which is essential for realistic generation since language influences phonetics, rhythm, intonation, and facial expressions. Speaking style is also shaped by individual differences, not only by language. Existing methods typically rely on either language-specific or speaker-specific conditioning, but not both, limiting their ability to model their interaction. We introduce Polyglot, a unified diffusion-based architecture for personalized multilingual SDFA. Our method uses transcript embeddings to encode language information and style embeddings extracted from reference facial sequences to capture individual speaking characteristics. Polyglot does not require predefined language or speaker labels, enabling generalization across languages and speakers through self-supervised learning. By jointly conditioning on language and style, it captures expressive traits such as rhythm, articulation, and habitual facial movements, producing temporally coherent and realistic animations. Experiments show improved performance in both monolingual and multilingual settings, providing a unified framework for modeling language and personal style in SDFA.