Learning Quantised Structure-Preserving Motion Representations for Dance Fingerprinting

arXiv cs.CV / 4/2/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces DANCEMATCH, an end-to-end framework for “dance fingerprinting,” enabling retrieval of semantically similar choreographies directly from raw video.
  • It addresses limitations of prior pose-sequence retrieval methods by replacing continuous embeddings with compact, discrete motion signatures that capture spatio-temporal structure and support efficient indexing.
  • DANCEMATCH combines Skeleton Motion Quantisation (SMQ) with Spatio-Temporal Transformers (STT) to quantise pose data (from Apple CoMotion) into a structured motion vocabulary.
  • It proposes a two-stage retrieval pipeline—DANCE RETRIEVAL ENGINE (DRE)—using a histogram-based, sub-linear index followed by re-ranking for more accurate matching.
  • The authors release DANCETYPESBENCHMARK, a pose-aligned dataset with quantised motion tokens to support reproducible research, and report strong cross-style retrieval and generalisation to unseen choreographies.

Abstract

We present DANCEMATCH, an end-to-end framework for motion-based dance retrieval, the task of identifying semantically similar choreographies directly from raw video, defined as DANCE FINGERPRINTING. While existing motion analysis and retrieval methods can compare pose sequences, they rely on continuous embeddings that are difficult to index, interpret, or scale. In contrast, DANCEMATCH constructs compact, discrete motion signatures that capture the spatio-temporal structure of dance while enabling efficient large-scale retrieval. Our system integrates Skeleton Motion Quantisation (SMQ) with Spatio-Temporal Transformers (STT) to encode human poses, extracted via Apple CoMotion, into a structured motion vocabulary. We further design DANCE RETRIEVAL ENGINE (DRE), which performs sub-linear retrieval using a histogram-based index followed by re-ranking for refined matching. To facilitate reproducible research, we release DANCETYPESBENCHMARK, a pose-aligned dataset annotated with quantised motion tokens. Experiments demonstrate robust retrieval across diverse dance styles and strong generalisation to unseen choreographies, establishing a foundation for scalable motion fingerprinting and quantitative choreographic analysis.