Beyond Perceptual Shortcuts: Causal-Inspired Debiasing Optimization for Generalizable Video Reasoning in Lightweight MLLMs

arXiv cs.CV / 5/5/2026

📰 NewsDeveloper Stack & InfrastructureIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that reinforcement learning can unintentionally limit reasoning quality in lightweight multimodal language models (MLLMs) by encouraging them to rely on perceptual shortcuts caused by dataset biases.
  • It proposes VideoThinker, a causal-inspired two-stage debiasing framework with Bias Aware Training to build an explicit “bias model,” followed by Causal Debiasing Policy Optimization (CDPO) to steer the main model away from the bias model’s flawed logic.
  • VideoThinker-R1 achieves new state-of-the-art results for efficient video reasoning, improving benchmark performance versus same-scale baselines with no supervised fine-tuning and reduced RL data usage.
  • In cross-scale evaluation, VideoThinker-R1 also outperforms a larger 7B model on multiple video reasoning benchmarks, indicating stronger generalization.
  • The authors provide code publicly, enabling others to reproduce and extend the approach for lightweight edge-deployable video reasoning systems.

Abstract

Although reinforcement learning (RL) has significantly advanced reasoning capabilities in large multimodal language models (MLLMs), its efficacy remains limited for lightweight models essential for edge deployments.To address this issue, we leverage causal analysis and experiment to reveal the underlying phenomenon of perceptual bias, demonstrating that RL-based fine-tuning compels lightweight models to preferentially adopt perceptual shortcuts induced by data biases, rather than developing genuine reasoning abilities.Motivated by this insight, we propose VideoThinker, a causal-inspired framework that cultivates robust reasoning in lightweight models through a two-stage debiasing process. First, the Bias Aware Training stage forges a dedicated "bias model" to embody these shortcut behaviors. Then, the Causal Debiasing Policy Optimization (CDPO) algorithm fine-tunes the primary model, employing an innovative repulsive objective to actively push it away from the bias model's flawed logic while simultaneously pulling it toward correct, generalizable solutions.Our model, VideoThinker-R1, establishes a new state-of-the-art in video reasoning efficiency. For same-scale comparison, requiring no Supervised Fine-Tuning (SFT) and using only 1 of the training data for RL, it surpasses VideoRFT-3B with a 3.2% average gain on widely-used benchmarks and a 7% lead on VideoMME. For cross-scale comparison, it outperforms the larger Video-UTR-7B model on multiple benchmarks, including a 2.1% gain on MVBench and a 3.8% gain on TempCompass. Code is available at https://github.com/falonss703/VideoThinker.