VSAS-BENCH: Real-Time Evaluation of Visual Streaming Assistant Models

arXiv cs.CV / 4/10/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • VSAS-BENCH is introduced as a new benchmark framework specifically for real-time visual streaming assistant (streaming VLM) evaluation, focusing on metrics beyond offline video understanding.
  • The benchmark includes temporally dense annotations (18,000+), diverse domains and task types, and provides standardized synchronous and asynchronous evaluation protocols.
  • It introduces metrics to separately measure proactiveness (response timeliness) and consistency (robustness of responses over time), enabling clearer analysis of streaming behavior.
  • Large-scale experiments evaluate accuracy–latency trade-offs across factors like memory buffer length, memory access policy, and input resolution, producing practical design insights.
  • The study shows conventional VLMs can be adapted to streaming without additional training and that adapted models (e.g., Qwen3-VL-4B) outperform prior streaming VLMs on VSAS-BENCH.

Abstract

Streaming vision-language models (VLMs) continuously generate responses given an instruction prompt and an online stream of input frames. This is a core mechanism for real-time visual assistants. Existing VLM frameworks predominantly assess models in offline settings. In contrast, the performance of a streaming VLM depends on additional metrics beyond pure video understanding, including proactiveness, which reflects the timeliness of the model's responses, and consistency, which captures the robustness of its responses over time. To address this limitation, we propose VSAS-Bench, a new framework and benchmark for Visual Streaming Assistants. In contrast to prior benchmarks that primarily employ single-turn question answering on video inputs, VSAS-Bench features temporally dense annotations with over 18,000 annotations across diverse input domains and task types. We introduce standardized synchronous and asynchronous evaluation protocols, along with metrics that isolate and measure distinct capabilities of streaming VLMs. Using this framework, we conduct large-scale evaluations of recent video and streaming VLMs, analyzing the accuracy-latency trade-off under key design factors such as memory buffer length, memory access policy, and input resolution, yielding several practical insights. Finally, we show empirically that conventional VLMs can be adapted to streaming settings without additional training, and demonstrate that these adapted models outperform recent streaming VLMs. For example, Qwen3-VL-4B surpasses Dispider, the best streaming VLM on our benchmark, by 3% under the asynchronous protocol. The benchmark and code will be available at https://github.com/apple/ml-vsas-bench.