PEARL: Personalized Streaming Video Understanding Model

arXiv cs.CV / 3/24/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • 提案論文は、リアルタイムに新しい概念を理解し記憶を更新する人の認知に対応するため、「パーソナライズされたストリーミング映像理解(PSVU)」という新しい課題設定を定義した。
  • 研究用ベンチマークとして、厳密なタイムスタンプ付きでフレーム単位と動画単位の2つの評価モードを備えた「PEARL-Bench」(132本の動画、2,173件の注釈)を導入した。
  • アノテーション多様性と品質を自動生成+人手検証のパイプラインで担保し、時刻に正確に反応できるかを測る点を強調している。
  • 課題に対し、学習を前提としないプラグ&プレイ手法「PEARL」を提案し、8つのオフライン/オンラインモデルでSOTA性能を達成したと報告している。
  • PEARLは異なる3つのアーキテクチャに適用しても一貫したPSVU改善が得られ、VLMのパーソナライズ化やストリーミング型AIアシスタント研究を促進することを狙っている。

Abstract

Human cognition of new concepts is inherently a streaming process: we continuously recognize new objects or identities and update our memories over time. However, current multimodal personalization methods are largely limited to static images or offline videos. This disconnects continuous visual input from instant real-world feedback, limiting their ability to provide the real-time, interactive personalized responses essential for future AI assistants. To bridge this gap, we first propose and formally define the novel task of Personalized Streaming Video Understanding (PSVU). To facilitate research in this new direction, we introduce PEARL-Bench, the first comprehensive benchmark designed specifically to evaluate this challenging setting. It evaluates a model's ability to respond to personalized concepts at exact timestamps under two modes: (1) Frame-level, focusing on a specific person or object in discrete frames, and (2) a novel Video-level, focusing on personalized actions unfolding across continuous frames. PEARL-Bench comprises 132 unique videos and 2,173 fine-grained annotations with precise timestamps. Concept diversity and annotation quality are strictly ensured through a combined pipeline of automated generation and human verification. To tackle this challenging new setting, we further propose PEARL, a plug-and-play, training-free strategy that serves as a strong baseline. Extensive evaluations across 8 offline and online models demonstrate that PEARL achieves state-of-the-art performance. Notably, it brings consistent PSVU improvements when applied to 3 distinct architectures, proving to be a highly effective and robust strategy. We hope this work advances vision-language model (VLM) personalization and inspires further research into streaming personalized AI assistants. Code is available at https://github.com/Yuanhong-Zheng/PEARL.