Modeling Subjective Urban Perception with Human Gaze
arXiv cs.CV / 5/4/2026
📰 NewsDeveloper Stack & InfrastructureIdeas & Deep AnalysisModels & Research
Key Points
- The paper highlights a gap in existing computational urban-perception work by noting that it often models subjective judgments directly from street-view images while largely ignoring the human perceptual (gaze) process behind those judgments.
- It introduces the Place Pulse-Gaze dataset, which pairs street-view images with synchronized eye-tracking recordings and individual perception labels to connect visual attention with subjective evaluation.
- The authors propose a Gaze-Guided Urban Perception Framework and evaluate three approaches: gaze-only modeling, gaze fused with explicit semantic scene representations, and gaze fused with implicit, richer visual representations.
- Experimental results show that gaze alone can meaningfully predict subjective urban perception, and that fusing gaze with scene representations improves prediction in both semantic and richer-visual settings.
- Overall, the findings argue for incorporating human perceptual mechanisms into urban scene understanding and suggest a path toward gaze-guided multimodal urban computing.
Related Articles

When Claims Freeze Because a Provider Record Drifted: The Case for Enrollment Repair Agents
Dev.to

The Cash Is Already Earned: Why Construction Pay Application Exceptions Fit an Agent Better Than SaaS
Dev.to

Why Ship-and-Debit Claim Recovery Is a Better Agent Wedge Than Another “AI Back Office” Tool
Dev.to
AI is getting better at doing things, but still bad at deciding what to do?
Reddit r/artificial

I Built an AI-Powered Chinese BaZi (八字) Fortune Teller — Here's What DeepSeek Revealed About Destiny
Dev.to