HuM-Eval: A Coarse-to-Fine Framework for Human-Centric Video Evaluation

arXiv cs.CV / 4/29/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The HuM-Eval paper introduces a human-centric video evaluation framework to better judge generated human-motion quality beyond coarse scene-level metrics.
  • It applies a coarse-to-fine pipeline: a vision-language model first assesses overall video quality, then 2D pose checks anatomical correctness and 3D motion analysis evaluates motion stability.
  • Experiments report an average human correlation of 58.2%, outperforming prior state-of-the-art baselines.
  • The authors also release HuM-Bench, a benchmark with 1,000 diverse prompts, and use it to evaluate existing text-to-video models, supporting progress toward next-generation human motion generation.

Abstract

Video generation models have developed rapidly in recent years, where generating natural human motion plays a pivotal role. However, accurately evaluating the quality of generated human motion video remains a significant challenge. Existing evaluation metrics primarily focus on global scene statistics, often overlooking fine-grained human details and consequently failing to align with human subjective preference. To bridge this gap, we propose HuM-Eval, a novel human-centric evaluation framework that adopts a coarse-to-fine strategy. Specifically, our framework first utilizes a Vision Language Model to perform a coarse assessment of global video quality. It then proceeds to a fine-grained analysis, using 2D pose to verify anatomical correctness and 3D human motion to evaluate motion stability. Extensive experiments demonstrate that HuM-Eval achieves an average human correlation of 58.2%, outperforming state-of-the-art baselines. Furthermore, we introduce HuM-Bench, a comprehensive benchmark comprising 1,000 diverse prompts, and conduct a detailed evaluation of existing text-to-video models, paving the way for next-generation human motion generation.