Modeling Subjective Urban Perception with Human Gaze

arXiv cs.CV / 5/4/2026

📰 NewsDeveloper Stack & InfrastructureIdeas & Deep AnalysisModels & Research

Key Points

  • The paper highlights a gap in existing computational urban-perception work by noting that it often models subjective judgments directly from street-view images while largely ignoring the human perceptual (gaze) process behind those judgments.
  • It introduces the Place Pulse-Gaze dataset, which pairs street-view images with synchronized eye-tracking recordings and individual perception labels to connect visual attention with subjective evaluation.
  • The authors propose a Gaze-Guided Urban Perception Framework and evaluate three approaches: gaze-only modeling, gaze fused with explicit semantic scene representations, and gaze fused with implicit, richer visual representations.
  • Experimental results show that gaze alone can meaningfully predict subjective urban perception, and that fusing gaze with scene representations improves prediction in both semantic and richer-visual settings.
  • Overall, the findings argue for incorporating human perceptual mechanisms into urban scene understanding and suggest a path toward gaze-guided multimodal urban computing.

Abstract

Urban perception describes how people subjectively evaluate urban environments, shaping how cities are experienced and understood. Existing computational approaches primarily model urban perception directly from street view images, but largely ignore the human perceptual process through which such judgments are formed. In this paper, we introduce Place Pulse-Gaze, an urban perception dataset that augments street view images with synchronized eye-tracking recordings and individual perception labels. Based on this dataset, we propose a Gaze-Guided Urban Perception Framework to study how gaze behavior contributes to the modeling of subjective urban perception. The framework systematically investigates three complementary settings: gaze-only modeling, gaze fusion with explicit semantic scene representations, and gaze fusion with implicit richer visual representations. Experiments show that gaze alone already carries useful predictive signals for subjective urban perception, and that integrating gaze with scene representations further improves prediction under both semantic and richer visual representations. Overall, our findings highlight the importance of incorporating human perceptual processes into urban scene understanding and open a direction for gaze-guided multimodal urban computing.

Modeling Subjective Urban Perception with Human Gaze | AI Navigate