EVA: Efficient Reinforcement Learning for End-to-End Video Agent

arXiv cs.CL / 3/25/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • EVA(Efficient Reinforcement Learning for End-to-End Video Agent)は、長い動画の冗長なフレームや時間的依存に起因するMLLMでの動画理解の非効率さを、強化学習で解決するエンドツーエンド動画エージェントの枠組みを提案しています。
  • EVAは「計画→知覚」の順に進める計画先行(summary-plan-action-reflection)を反復し、必要な部分だけを“いつ・何を・どう”観るかを自律的に意思決定します。
  • 学習はSFT(教師あり微調整)からKTO(Kahneman-Tversky Optimization)、GRPO(Generalized Reward Policy Optimization)へと進む3段階パイプラインで、模倣学習と強化学習を橋渡しする設計になっています。
  • 6つの動画理解ベンチマークで評価し、一般的なMLLMベースラインに対して6-12%の改善、既存の適応型エージェントに対してさらに1-3%上乗せする結果を報告しています。

Abstract

Video understanding with multimodal large language models (MLLMs) remains challenging due to the long token sequences of videos, which contain extensive temporal dependencies and redundant frames. Existing approaches typically treat MLLMs as passive recognizers, processing entire videos or uniformly sampled frames without adaptive reasoning. Recent agent-based methods introduce external tools, yet still depend on manually designed workflows and perception-first strategies, resulting in inefficiency on long videos. We present EVA, an Efficient Reinforcement Learning framework for End-to-End Video Agent, which enables planning-before-perception through iterative summary-plan-action-reflection reasoning. EVA autonomously decides what to watch, when to watch, and how to watch, achieving query-driven and efficient video understanding. To train such agents, we design a simple yet effective three-stage learning pipeline - comprising supervised fine-tuning (SFT), Kahneman-Tversky Optimization (KTO), and Generalized Reward Policy Optimization (GRPO) - that bridges supervised imitation and reinforcement learning. We further construct high-quality datasets for each stage, supporting stable and reproducible training. We evaluate EVA on six video understanding benchmarks, demonstrating its comprehensive capabilities. Compared with existing baselines, EVA achieves a substantial improvement of 6-12% over general MLLM baselines and a further 1-3% gain over prior adaptive agent methods. Our code and model are available at https://github.com/wangruohui/EfficientVideoAgent.