AI Navigate

Symphony: A Cognitively-Inspired Multi-Agent System for Long-Video Understanding

arXiv cs.CV / 3/19/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • Symphony proposes a cognitively-inspired multi-agent system to improve long-video understanding by decomposing LVU into fine-grained subtasks and enabling reflection-enhanced deep reasoning collaboration.
  • It introduces a VLM-based grounding approach to analyze LVU tasks and assess the relevance of video segments to identify complex problems with long temporal spans.
  • The method aims to overcome limitations of simple task decomposition and embedding-based retrieval that risk losing key information in long contexts.
  • Experiments show state-of-the-art performance on LVBench, LongVideoBench, VideoMME, and MLVU, with a 5.0% improvement on LVBench, and the code is available on GitHub.

Abstract

Despite rapid developments and widespread applications of MLLM agents, they still struggle with long-form video understanding (LVU) tasks, which are characterized by high information density and extended temporal spans. Recent research on LVU agents demonstrates that simple task decomposition and collaboration mechanisms are insufficient for long-chain reasoning tasks. Moreover, directly reducing the time context through embedding-based retrieval may lose key information of complex problems. In this paper, we propose Symphony, a multi-agent system, to alleviate these limitations. By emulating human cognition patterns, Symphony decomposes LVU into fine-grained subtasks and incorporates a deep reasoning collaboration mechanism enhanced by reflection, effectively improving the reasoning capability. Additionally, Symphony provides a VLM-based grounding approach to analyze LVU tasks and assess the relevance of video segments, which significantly enhances the ability to locate complex problems with implicit intentions and large temporal spans. Experimental results show that Symphony achieves state-of-the-art performance on LVBench, LongVideoBench, VideoMME, and MLVU, with a 5.0% improvement over the prior state-of-the-art method on LVBench. Code is available at https://github.com/Haiyang0226/Symphony.