MASRA: MLLM-Assisted Semantic-Relational Consistent Alignment for Video Temporal Grounding

arXiv cs.CV / 5/6/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • MASRA is a training-time framework for Video Temporal Grounding (VTG) that tackles the cross-modal semantic gap and improves how video moments are aligned to queries.
  • It uses an MLLM during training to generate textual priors in two forms—event-level descriptions with temporal spans and clip-level captions—and then performs two MLLM-assisted alignments.
  • ESTA strengthens span-level separability by aligning temporal context with event semantics, while LRCA improves temporal consistency by matching a relation matrix from captions to the model’s temporal feature similarity matrix.
  • MASRA adds modules (semantic-guided enhancement and second-order relational attention) plus Decoupled Alignment Interaction (DAI) with a context-aware codebook to reduce query-irrelevant semantics.
  • The approach is reported to outperform prior methods in extensive experiments, and the MLLM is not used at inference, improving deployability.

Abstract

Video Temporal Grounding (VTG) faces a cross-modal semantic gap that often leads to background features being incorrectly aligned with the query, while directly matching the query to moments results in insufficient discriminability and consistency of temporal semantics. To address this issue, we propose MLLM-Assisted Semantic-Relational Consistent Alignment (MASRA), a training-time MLLM-based optimization framework for VTG. MASRA leverages an MLLM during training to produce two forms of textual priors, namely event-level descriptions with temporal spans and clip-level captions, and instantiates two MLLM-assisted alignments. Event Semantic Temporal Alignment (ESTA) aligns temporal context with event semantics to explicitly strengthen the correspondence between semantics and temporal events and improve span-level separability. Local Relational Consistency Alignment (LRCA) constructs a textual relation matrix derived from clip-level captions and aligns it with the temporal feature similarity matrix in the model, enhancing temporal consistency while capturing local structural information. MASRA includes two simple supporting modules, semantic-guided enhancement and second-order relational attention, to better utilize the learned semantic context and relational structure. Moreover, we introduce Decoupled Alignment Interaction (DAI) with a context-aware codebook to adaptively absorb query-irrelevant semantics and alleviate the cross-modal gap. The MLLM is only invoked during training and is not used at inference. Extensive experiments show that MASRA outperforms existing methods, and ablation studies validate its effectiveness.