Gaze-Regularized VLMs for Ego-Centric Behavior Understanding

arXiv cs.CV / 3/25/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes a gaze-regularized training framework that injects eye-gaze information (fixations and saccades) into Vision Language Models for egocentric behavior understanding.
  • It uses gaze-based queries and a gaze-regularization mechanism so the model’s attention aligns with human attention patterns rather than relying on vision-only inputs.
  • The authors run extensive experiments comparing multiple strategies for incorporating gaze data into the VLM architecture.
  • Results show nearly a 13% improvement in semantic scores over baseline models that do not use gaze information, enabling better future event prediction with detailed action descriptions.
  • The work is positioned as a foundation for leveraging human gaze signals to improve VLM predictive capability in applications that require robust understanding of future actions.

Abstract

Eye gaze, encompassing fixations and saccades, provides critical insights into human intentions and future actions. This study introduces a gaze-regularized framework that enhances Vision Language Models (VLMs) for egocentric behavior understanding. Unlike existing methods that rely solely on visual data and overlook gaze information, our approach directly incorporates gaze information into the VLM architecture during training. By generating gaze-based queries, the model dynamically focuses on gaze-highlighted regions, while a gaze-regularization mechanism ensures the alignment of model attention with human attention patterns. To better understand how gaze can be effectively integrated into VLMs, we conducted extensive experiments exploring various strategies for incorporating gaze data. These innovations enable the prediction of future events with detailed action descriptions. Experimental results demonstrate a nearly 13 % improvement in semantic scores compared to baseline models not leveraging gaze data, highlighting the effectiveness of our approach. This work establishes a foundation for leveraging the human gaze in VLMs, significantly boosting their predictive capabilities in applications requiring accurate and robust future event prediction.