Contextual inference from single objects in Vision-Language models

arXiv cs.CV / 3/31/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies how vision-language models infer scene context from single objects by testing fine-grained scene category and coarse indoor-vs-outdoor context on masked backgrounds.
  • Experiments show above-chance contextual inference at both levels, with performance influenced by object properties similarly to human scene categorization.
  • The model’s object identity, scene, and superordinate context predictions are partially dissociable, meaning strong accuracy in one level does not imply accuracy in the others and coupling varies by model.
  • Mechanistic analysis indicates that object representations that stay stable after removing background are most predictive of successful contextual inference.
  • It finds different internal grounding for scene vs. superordinate schemas: scene identity is encoded broadly in image tokens across the network, while superordinate information appears only late or not reliably, suggesting complex organization beyond end accuracy.

Abstract

How much scene context a single object carries is a well-studied question in human scene perception, yet how this capacity is organized in vision-language models (VLMs) remains poorly understood, with direct implications for the robustness of these models. We investigate this question through a systematic behavioral and mechanistic analysis of contextual inference from single objects. Presenting VLMs with single objects on masked backgrounds, we probe their ability to infer both fine-grained scene category and coarse superordinate context (indoor vs. outdoor). We found that single objects support above-chance inference at both levels, with performance modulated by the same object properties that predict human scene categorization. Object identity, scene, and superordinate predictions are partially dissociable: accurate inference at one level neither requires nor guarantees accurate inference at the others, and the degree of coupling differs markedly across models. Mechanistically, object representations that remain stable when background context is removed are more predictive of successful contextual inference. Scene and superordinate schemas are grounded in fundamentally different ways: scene identity is encoded in image tokens throughout the network, while superordinate information emerges only late or not at all. Together, these results reveal that the organization of contextual inference in VLMs is more complex than accuracy alone suggests, with behavioral and mechanistic signatures