SaSaSaSa2VA: 2nd Place of the 5th PVUW MeViS-Text Track

arXiv cs.CV / 3/31/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper SaSaSaSa2VA targets referring video object segmentation (RVOS), arguing that existing approaches rely too heavily on static textual cues and thus extend the setting toward motion-centric expressions.
  • It builds on Sa2VA by increasing input frames and using [SEG] tokens, then adds a simple target existence-aware verification mechanism inspired by the need to verify whether targets exist before/while segmenting.
  • The authors report a final score of 89.19 at the 5th PVUW Challenge (MeViS-Text Track), where the method won 2nd place.
  • Quantitative results and ablation studies indicate that the existence-aware verification strategy is sufficient to unlock strong performance specifically on motion-centric referring tasks.
  • The work positions MeViS benchmark improvements (referring & reasoning motion expressions plus no-target queries) as a key testbed for evaluating robustness beyond text-only grounding.

Abstract

Referring video object segmentation (RVOS) commonly grounds targets in videos based on static textual cues. MeViS benchmark extends this by incorporating motion-centric expressions (referring & reasoning motion expressions) and introducing no-target queries. Extending SaSaSa2VA, where increased input frames and [SEG] tokens already strengthen the Sa2VA backbone, we adopt a simple yet effective target existence-aware verification mechanism, leading to Still Awesome SaSaSa2VA (SaSaSaSa2VA). Despite its simplicity, the method achieves a final score of 89.19 in the 5th PVUW Challenge (MeViS-Text Track), securing 2nd place. Both quantitative results and ablations suggest that this existence-aware verification strategy is sufficient to unlock strong performance on motion-centric referring tasks.