How Should Video LLMs Output Time? An Analysis of Efficient Temporal Grounding Paradigms

arXiv cs.CV / 4/13/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper analyzes how different output formats for Video Temporal Grounding (VTG)—Text Numeral Generation, Temporal Token Generation, and Continuous Temporal Decoding—affect both localization accuracy and computational efficiency.
  • It runs a controlled comparison using the same compact VLM backbones (SmolVLM2, FastVLM, Molmo2), consistent datasets, and LoRA fine-tuning protocols to isolate the impact of output design.
  • Evaluations on Charades-STA, QVHighlights, and YouCook2 measure grounding quality alongside system-level metrics such as inference latency, training throughput, and parameter overhead.
  • The findings indicate that output formulation can change the efficiency–accuracy trade-off significantly, largely independent of model scale.
  • Continuous Temporal Decoding is reported to yield the best Pareto-front performance, providing robust localization with minimal latency overhead and supporting deployment on resource-constrained edge devices.

Abstract

While Multimodal Large Language Models (MLLMs) have advanced Video Temporal Grounding (VTG), existing methods often couple output paradigms with different backbones, datasets, and training protocols. This makes it challenging to isolate the specific impact of the output design. Additionally, as VTG systems are increasingly considered for resource-constrained edge deployment, the trade-off between output formulation and system-level efficiency requires systematic investigation. In this paper, we present a controlled empirical study comparing three dominant VTG output paradigms: Text Numeral Generation, Temporal Token Generation, and Continuous Temporal Decoding. We evaluate these paradigms across identical compact VLMs (SmolVLM2, FastVLM, and Molmo2) using consistent datasets and LoRA fine-tuning protocols. Evaluations on Charades-STA, QVHighlights, and YouCook2 measure both localization accuracy and system efficiency, including inference latency, training throughput, and parameter overhead. Our results demonstrate that the choice of output formulation significantly affects both grounding accuracy and computational cost, independent of model scale. Specifically, the continuous distribution paradigm consistently achieves the most favorable efficiency-accuracy trade-off on the Pareto frontier, delivering robust localization with minimal latency overhead. These findings provide objective empirical guidelines for designing efficient, deployment-ready VTG systems.