M3T: Discrete Multi-Modal Motion Tokens for Sign Language Production

arXiv cs.CV / 3/26/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that sign language production must generate non-manual features (e.g., mouthings, eyebrow raises, gaze, head movement) because these are grammatically obligatory and not recoverable from hand motion alone.
  • It introduces SMPL-FX to combine FLAME facial expressiveness with the SMPL-X body, and uses modality-specific Finite Scalar Quantization VAEs to discretize body, hands, and face representations.
  • M3T is an autoregressive transformer trained over the resulting multi-modal motion token vocabulary, with an auxiliary translation objective to encourage semantically grounded embeddings.
  • Experiments on How2Sign, CSL-Daily, and Phoenix14T show state-of-the-art sign language production quality, and on NMFs-CSL it attains 58.3% accuracy vs. 49.0% for the strongest comparable pose baseline.

Abstract

Sign language production requires more than hand motion generation. Non-manual features, including mouthings, eyebrow raises, gaze, and head movements, are grammatically obligatory and cannot be recovered from manual articulators alone. Existing 3D production systems face two barriers to integrating them: the standard body model provides a facial space too low-dimensional to encode these articulations, and when richer representations are adopted, standard discrete tokenization suffers from codebook collapse, leaving most of the expression space unreachable. We propose SMPL-FX, which couples FLAME's rich expression space with the SMPL-X body, and tokenize the resulting representation with modality-specific Finite Scalar Quantization VAEs for body, hands, and face. M3T is an autoregressive transformer trained on this multi-modal motion vocabulary, with an auxiliary translation objective that encourages semantically grounded embeddings. Across three standard benchmarks (How2Sign, CSL-Daily, Phoenix14T) M3T achieves state-of-the-art sign language production quality, and on NMFs-CSL, where signs are distinguishable only by non-manual features, reaches 58.3% accuracy against 49.0% for the strongest comparable pose baseline.