ProCap: Projection-Aware Captioning for Spatial Augmented Reality

arXiv cs.CV / 4/2/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • ProCap is proposed to solve virtual–physical semantic ambiguity in Spatial Augmented Reality (SAR), where projectors can cause vision-language models to confuse projected content with the real scene.
  • The framework uses a two-stage pipeline: automated segmentation to decouple virtual and physical layers, followed by region-aware retrieval to reduce projection-distortion-related context ambiguity.
  • The paper introduces RGBP (RGB + Projections), a large-scale SAR semantic benchmark with 65 physical scenes, 180,000+ projections, and dense annotations that separately capture decoupled scene/projection semantics.
  • A dual-captioning evaluation protocol is defined with task-specific tokens to independently assess descriptions of the physical scene versus the projected content.
  • The authors report that ProCap yields a more robust semantic foundation for intelligent SAR interaction and release code, pre-trained models, and the dataset.

Abstract

Spatial augmented reality (SAR) directly projects digital content onto physical scenes using projectors, creating immersive experience without head-mounted displays. However, for SAR to support intelligent interaction, such as reasoning about the scene or answering user queries, it must semantically distinguish between the physical scene and the projected content. Standard Vision Language Models (VLMs) struggle with this virtual-physical ambiguity, often confusing the two contexts. To address this issue, we introduce ProCap, a novel framework that explicitly decouples projected content from physical scenes. ProCap employs a two-stage pipeline: first it visually isolates virtual and physical layers via automated segmentation; then it uses region-aware retrieval to avoid ambiguous semantic context due to projection distortion. To support this, we present RGBP (RGB + Projections), the first large-scale SAR semantic benchmark dataset, featuring 65 diverse physical scenes and over 180,000 projections with dense, decoupled annotations. Finally, we establish a dual-captioning evaluation protocol using task-specific tokens to assess physical scene and projection descriptions independently. Our experiments show that ProCap provides a robust semantic foundation for future SAR research. The source code, pre-trained models and the RGBP dataset are available on the project page: https://ZimoCao.github.io/ProCap/.