Physion-Eval: Evaluating Physical Realism in Generated Video via Human Reasoning

arXiv cs.CV / 3/23/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • Physion-Eval introduces a large-scale benchmark that uses expert human reasoning to diagnose physical realism failures in videos generated by five state-of-the-art models across egocentric and exocentric views, with 10,990 reasoning traces spanning 22 fine-grained categories.
  • Each generated video is paired with a corresponding real-world reference and annotated with temporally localized glitches, structured failure categories, and natural-language explanations of the violated physical behaviors.
  • The study reveals that in physics-critical scenarios, 83.3% of exocentric and 93.5% of egocentric generated videos exhibit at least one human-identifiable physical glitch.
  • The benchmark addresses limitations of automated metrics and rough judgments by focusing on human reasoning about physical constraints, aiming to guide the development of physics-grounded video generation.
  • The Physion-Eval dataset is publicly available on HuggingFace, enabling researchers to benchmark and advance physically realistic video generation.

Abstract

Video generation models are increasingly used as world simulators for storytelling, simulation, and embodied AI. As these models advance, a key question arises: do generated videos obey the physical laws of the real world? Existing evaluations largely rely on automated metrics or coarse human judgments such as preferences or rubric-based checks. While useful for assessing perceptual quality, these methods provide limited insight into when and why generated dynamics violate real-world physical constraints. We introduce Physion-Eval, a large-scale benchmark of expert human reasoning for diagnosing physical realism failures in videos generated by five state-of-the-art models across egocentric and exocentric views, containing 10,990 expert reasoning traces spanning 22 fine-grained physical categories. Each generated video is derived from a corresponding real-world reference video depicting a clear physical process, and annotated with temporally localized glitches, structured failure categories, and natural-language explanations of the violated physical behavior. Using this dataset, we reveal a striking limitation of current video generation models: in physics-critical scenarios, 83.3% of exocentric and 93.5% of egocentric generated videos exhibit at least one human-identifiable physical glitch. We hope Physion-Eval will set a new standard for physical realism evaluation and guide the development of physics-grounded video generation. The benchmark is publicly available at https://huggingface.co/datasets/PhysionLabs/Physion-Eval.