Watch Before You Answer: Learning from Visually Grounded Post-Training

arXiv cs.CL / 4/8/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that current vision-language model (VLM) video understanding is weaker than expected because many long-video benchmarks (and even post-training datasets) contain 40–60% of questions answerable from text cues alone.
  • It reports that this “linguistic shortcut” problem can undermine the effectiveness of post-training aimed at improving visual grounding, since models may learn to rely on language rather than video content.
  • To address this, the authors propose VidGround, a data curation/post-training method that keeps only questions that are truly visually grounded while removing linguistic biases.
  • When combined with RL-based post-training, VidGround improves video understanding by up to 6.2 points compared with training on the full (biased) dataset, while using only 69.1% of the original post-training data.
  • The study concludes that data curation quality is a key bottleneck, with simpler curation plus a straightforward post-training algorithm outperforming several more complex approaches.

Abstract

It is critical for vision-language models (VLMs) to comprehensively understand visual, temporal, and textual cues. However, despite rapid progress in multimodal modeling, video understanding performance still lags behind text-based reasoning. In this work, we find that progress is even worse than previously assumed: commonly reported long video understanding benchmarks contain 40-60% of questions that can be answered using text cues alone. Furthermore, we find that these issues are also pervasive in widely used post-training datasets, potentially undercutting the ability of post-training to improve VLM video understanding performance. Guided by this observation, we introduce VidGround as a simple yet effective solution: using only the actual visually grounded questions without any linguistic biases for post-training. When used in tandem with RL-based post-training algorithms, this simple technique improves performance by up to 6.2 points relative to using the full dataset, while using only 69.1% of the original post-training data. Moreover, we show that data curation with a simple post-training algorithm outperforms several more complex post-training techniques, highlighting that data quality is a major bottleneck for improving video understanding in VLMs. These results underscore the importance of curating post-training data and evaluation benchmarks that truly require visual grounding to advance the development of more capable VLMs. Project page: http://vidground.etuagi.com.