Automated Detection of Mutual Gaze and Joint Attention in Dual-Camera Settings via Dual-Stream Transformers

arXiv cs.CV / 5/1/2026

📰 NewsSignals & Early TrendsTools & Practical UsageModels & Research

Key Points

  • The paper proposes an efficient dual-stream Transformer model to automatically detect mutual gaze and joint attention from synchronized dual-camera recordings in multi-camera lab settings.
  • It builds on frozen gaze-aware backbones (GazeLLE) to capture strong visual priors and uses a custom token-fusion mechanism to model spatial and semantic relations between interacting subjects.
  • Experiments on an ecologically valid caregiver–infant interaction dataset show the method performs well and significantly better than both a convolutional baseline and a state-of-the-art multimodal LLM.
  • The authors open-source the model and pre-trained weights to enable behavioral scientists to fine-tune the system for different laboratory environments, reducing reliance on labor-intensive manual coding.
  • Overall, the work bridges computational modeling and applied interaction research by offering a scalable pipeline for behavioral measurement.

Abstract

Analyzing mutual gaze (MG) and joint attention (JA) is critical in developmental psychology but traditionally relies on labor-intensive manual coding. Automating this process in multi-camera laboratory settings is computationally challenging due to complex cross-camera relational dynamics. In this paper, we propose a highly efficient dual-stream Transformer architecture for detecting MG and JA from synchronized dual-camera recordings. Our approach leverages frozen gaze-aware backbones (GazeLLE) to extract rich visual priors, combined with a custom token fusion mechanism to map the spatial and semantic relationships between interacting dyads. Evaluated on an ecologically valid dataset of caregiver-infant interactions, our model exhibits good performance, significantly outperforming both a convolutional baseline and a state-of-the-art multimodal Large Language Model (LLM). By open-sourcing our model and pre-trained weights, we provide behavioral scientists with a scalable tool that can be fine-tuned to diverse laboratory environments, effectively bridging the gap between computational modeling and applied interaction research.