Where are they looking in the operating room?
arXiv cs.CV / 4/23/2026
📰 NewsDeveloper Stack & InfrastructureIdeas & Deep AnalysisModels & Research
Key Points
- The paper introduces “gaze-following” to the operating room, aiming to infer where surgical staff are looking to improve understanding of attention in high-stakes surgical workflows.
- It extends existing surgical video datasets by adding gaze-following annotations to 4D-OR and gaze-following plus new team-communication activity labels to Team-OR.
- The authors propose gaze-based models for three downstream tasks: clinical role prediction and surgical phase recognition using gaze heatmaps, and team communication detection using self-supervised spatial-temporal gaze features.
- On the 4D-OR and Team-OR benchmarks, the method reaches state-of-the-art results, achieving F1 scores of 0.92 (role prediction) and 0.95 (phase recognition).
- For team communication detection, the approach surpasses prior baselines by more than 30%, indicating substantial gains in recognizing coordination signals from gaze.
Related Articles

Big Tech firms are accelerating AI investments and integration, while regulators and companies focus on safety and responsible adoption.
Dev.to

Trajectory Forecasts in Unknown Environments Conditioned on Grid-Based Plans
Dev.to

Elevating Austria: Google invests in its first data center in the Alps.
Google Blog

Why use an AI gateway at all?
Dev.to

OpenAI Just Named It Workspace Agents. We Open-Sourced Our Lark Version Six Months Ago
Dev.to